var/home/core/zuul-output/0000755000175000017500000000000015067150363014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067204325015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000007146557615067204315017725 0ustar rootrootOct 01 07:05:41 crc systemd[1]: Starting Kubernetes Kubelet... Oct 01 07:05:42 crc restorecon[4788]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:42 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 07:05:43 crc restorecon[4788]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 01 07:05:43 crc kubenswrapper[4837]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.546269 4837 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554653 4837 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554715 4837 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554728 4837 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554737 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554747 4837 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554757 4837 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554765 4837 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554774 4837 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554793 4837 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554804 4837 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554813 4837 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554821 4837 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554830 4837 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554838 4837 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554846 4837 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554855 4837 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554863 4837 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554877 4837 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554886 4837 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554898 4837 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554908 4837 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554918 4837 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554926 4837 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554935 4837 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554944 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554953 4837 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554964 4837 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554975 4837 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554984 4837 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.554995 4837 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555005 4837 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555014 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555024 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555036 4837 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555048 4837 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555058 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555067 4837 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555076 4837 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555084 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555093 4837 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555101 4837 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555109 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555118 4837 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555126 4837 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555135 4837 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555143 4837 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555151 4837 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555160 4837 feature_gate.go:330] unrecognized feature gate: Example Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555173 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555182 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555191 4837 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555199 4837 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555207 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555216 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555224 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555233 4837 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555243 4837 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555252 4837 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555260 4837 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555272 4837 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555283 4837 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555291 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555300 4837 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555308 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555316 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555324 4837 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555332 4837 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555342 4837 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555350 4837 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555359 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.555367 4837 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556275 4837 flags.go:64] FLAG: --address="0.0.0.0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556301 4837 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556319 4837 flags.go:64] FLAG: --anonymous-auth="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556332 4837 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556345 4837 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556355 4837 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556367 4837 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556379 4837 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556389 4837 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556399 4837 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556410 4837 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556420 4837 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556430 4837 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556440 4837 flags.go:64] FLAG: --cgroup-root="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556449 4837 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556459 4837 flags.go:64] FLAG: --client-ca-file="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556469 4837 flags.go:64] FLAG: --cloud-config="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556479 4837 flags.go:64] FLAG: --cloud-provider="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556489 4837 flags.go:64] FLAG: --cluster-dns="[]" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556499 4837 flags.go:64] FLAG: --cluster-domain="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556509 4837 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556519 4837 flags.go:64] FLAG: --config-dir="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556529 4837 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556539 4837 flags.go:64] FLAG: --container-log-max-files="5" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556550 4837 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556560 4837 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556572 4837 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556582 4837 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556592 4837 flags.go:64] FLAG: --contention-profiling="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556602 4837 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556612 4837 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556622 4837 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556633 4837 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556645 4837 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556654 4837 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556664 4837 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556674 4837 flags.go:64] FLAG: --enable-load-reader="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556683 4837 flags.go:64] FLAG: --enable-server="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556732 4837 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556745 4837 flags.go:64] FLAG: --event-burst="100" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556755 4837 flags.go:64] FLAG: --event-qps="50" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556765 4837 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556775 4837 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556785 4837 flags.go:64] FLAG: --eviction-hard="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556796 4837 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556806 4837 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556816 4837 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556827 4837 flags.go:64] FLAG: --eviction-soft="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556837 4837 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556847 4837 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556858 4837 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556868 4837 flags.go:64] FLAG: --experimental-mounter-path="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556878 4837 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556887 4837 flags.go:64] FLAG: --fail-swap-on="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556896 4837 flags.go:64] FLAG: --feature-gates="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556908 4837 flags.go:64] FLAG: --file-check-frequency="20s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556918 4837 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556928 4837 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556938 4837 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556947 4837 flags.go:64] FLAG: --healthz-port="10248" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556957 4837 flags.go:64] FLAG: --help="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556969 4837 flags.go:64] FLAG: --hostname-override="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556979 4837 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556989 4837 flags.go:64] FLAG: --http-check-frequency="20s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.556999 4837 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557008 4837 flags.go:64] FLAG: --image-credential-provider-config="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557017 4837 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557027 4837 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557038 4837 flags.go:64] FLAG: --image-service-endpoint="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557047 4837 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557056 4837 flags.go:64] FLAG: --kube-api-burst="100" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557066 4837 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557076 4837 flags.go:64] FLAG: --kube-api-qps="50" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557085 4837 flags.go:64] FLAG: --kube-reserved="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557095 4837 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557104 4837 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557114 4837 flags.go:64] FLAG: --kubelet-cgroups="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557124 4837 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557133 4837 flags.go:64] FLAG: --lock-file="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557143 4837 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557152 4837 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557162 4837 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557176 4837 flags.go:64] FLAG: --log-json-split-stream="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557186 4837 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557196 4837 flags.go:64] FLAG: --log-text-split-stream="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557205 4837 flags.go:64] FLAG: --logging-format="text" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557214 4837 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557225 4837 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557235 4837 flags.go:64] FLAG: --manifest-url="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557244 4837 flags.go:64] FLAG: --manifest-url-header="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557264 4837 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557275 4837 flags.go:64] FLAG: --max-open-files="1000000" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557287 4837 flags.go:64] FLAG: --max-pods="110" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557297 4837 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557308 4837 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557318 4837 flags.go:64] FLAG: --memory-manager-policy="None" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557328 4837 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557338 4837 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557347 4837 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557357 4837 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557378 4837 flags.go:64] FLAG: --node-status-max-images="50" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557387 4837 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557397 4837 flags.go:64] FLAG: --oom-score-adj="-999" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557407 4837 flags.go:64] FLAG: --pod-cidr="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557418 4837 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557433 4837 flags.go:64] FLAG: --pod-manifest-path="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557442 4837 flags.go:64] FLAG: --pod-max-pids="-1" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557452 4837 flags.go:64] FLAG: --pods-per-core="0" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557462 4837 flags.go:64] FLAG: --port="10250" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557472 4837 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557482 4837 flags.go:64] FLAG: --provider-id="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557491 4837 flags.go:64] FLAG: --qos-reserved="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557501 4837 flags.go:64] FLAG: --read-only-port="10255" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557510 4837 flags.go:64] FLAG: --register-node="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557523 4837 flags.go:64] FLAG: --register-schedulable="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557533 4837 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557548 4837 flags.go:64] FLAG: --registry-burst="10" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557558 4837 flags.go:64] FLAG: --registry-qps="5" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557567 4837 flags.go:64] FLAG: --reserved-cpus="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557577 4837 flags.go:64] FLAG: --reserved-memory="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557590 4837 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557600 4837 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557610 4837 flags.go:64] FLAG: --rotate-certificates="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557620 4837 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557630 4837 flags.go:64] FLAG: --runonce="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557639 4837 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557649 4837 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557659 4837 flags.go:64] FLAG: --seccomp-default="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557668 4837 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557678 4837 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557713 4837 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557724 4837 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557734 4837 flags.go:64] FLAG: --storage-driver-password="root" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557743 4837 flags.go:64] FLAG: --storage-driver-secure="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557753 4837 flags.go:64] FLAG: --storage-driver-table="stats" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557762 4837 flags.go:64] FLAG: --storage-driver-user="root" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557772 4837 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557782 4837 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557791 4837 flags.go:64] FLAG: --system-cgroups="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557801 4837 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557817 4837 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557826 4837 flags.go:64] FLAG: --tls-cert-file="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557836 4837 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557847 4837 flags.go:64] FLAG: --tls-min-version="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557856 4837 flags.go:64] FLAG: --tls-private-key-file="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557866 4837 flags.go:64] FLAG: --topology-manager-policy="none" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557877 4837 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557888 4837 flags.go:64] FLAG: --topology-manager-scope="container" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557898 4837 flags.go:64] FLAG: --v="2" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557911 4837 flags.go:64] FLAG: --version="false" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557923 4837 flags.go:64] FLAG: --vmodule="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557934 4837 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.557945 4837 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558213 4837 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558226 4837 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558237 4837 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558247 4837 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558257 4837 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558266 4837 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558274 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558283 4837 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558292 4837 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558301 4837 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558310 4837 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558321 4837 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558333 4837 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558342 4837 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558353 4837 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558362 4837 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558373 4837 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558384 4837 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558394 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558403 4837 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558411 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558420 4837 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558429 4837 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558439 4837 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558448 4837 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558457 4837 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558465 4837 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558473 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558482 4837 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558491 4837 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558499 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558507 4837 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558516 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558525 4837 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558533 4837 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558541 4837 feature_gate.go:330] unrecognized feature gate: Example Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558550 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558558 4837 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558567 4837 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558576 4837 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558585 4837 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558593 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558602 4837 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558610 4837 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558619 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558629 4837 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558646 4837 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558654 4837 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558663 4837 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558671 4837 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558680 4837 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558713 4837 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558721 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558729 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558738 4837 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558747 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558758 4837 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558769 4837 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558779 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558792 4837 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558803 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558814 4837 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558824 4837 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558834 4837 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558844 4837 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558855 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558866 4837 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558877 4837 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558891 4837 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558902 4837 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.558911 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.558940 4837 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.573049 4837 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.573106 4837 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573244 4837 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573259 4837 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573269 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573279 4837 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573288 4837 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573297 4837 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573306 4837 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573315 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573325 4837 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573333 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573342 4837 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573350 4837 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573358 4837 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573368 4837 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573378 4837 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573388 4837 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573400 4837 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573413 4837 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573424 4837 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573433 4837 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573442 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573452 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573461 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573469 4837 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573478 4837 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573486 4837 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573496 4837 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573505 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573513 4837 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573522 4837 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573530 4837 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573539 4837 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573547 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573555 4837 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573565 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573577 4837 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573587 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573597 4837 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573605 4837 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573613 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573622 4837 feature_gate.go:330] unrecognized feature gate: Example Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573630 4837 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573642 4837 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573655 4837 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573665 4837 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573675 4837 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573685 4837 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573726 4837 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573741 4837 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573752 4837 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573762 4837 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573771 4837 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573782 4837 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573791 4837 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573800 4837 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573809 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573818 4837 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573827 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573836 4837 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573846 4837 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573855 4837 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573864 4837 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573873 4837 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573882 4837 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573891 4837 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573899 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573908 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573916 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573925 4837 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573933 4837 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.573943 4837 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.573957 4837 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574229 4837 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574243 4837 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574253 4837 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574264 4837 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574277 4837 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574287 4837 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574297 4837 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574307 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574317 4837 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574326 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574335 4837 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574345 4837 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574356 4837 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574366 4837 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574376 4837 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574385 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574395 4837 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574404 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574413 4837 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574421 4837 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574429 4837 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574438 4837 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574446 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574455 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574463 4837 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574471 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574479 4837 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574488 4837 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574496 4837 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574505 4837 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574513 4837 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574521 4837 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574531 4837 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574540 4837 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574549 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574558 4837 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574566 4837 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574574 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574583 4837 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574591 4837 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574599 4837 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574609 4837 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574618 4837 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574626 4837 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574634 4837 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574643 4837 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574651 4837 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574660 4837 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574668 4837 feature_gate.go:330] unrecognized feature gate: Example Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574677 4837 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574685 4837 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574743 4837 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574751 4837 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574760 4837 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574771 4837 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574782 4837 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574792 4837 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574802 4837 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574811 4837 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574821 4837 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574831 4837 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574840 4837 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574849 4837 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574857 4837 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574866 4837 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574874 4837 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574883 4837 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574891 4837 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574899 4837 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574908 4837 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.574920 4837 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.574933 4837 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.576840 4837 server.go:940] "Client rotation is on, will bootstrap in background" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.584281 4837 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.584433 4837 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.586323 4837 server.go:997] "Starting client certificate rotation" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.586378 4837 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.586608 4837 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-19 16:24:26.719436226 +0000 UTC Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.586721 4837 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1905h18m43.132718177s for next certificate rotation Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.616431 4837 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.622383 4837 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.643597 4837 log.go:25] "Validated CRI v1 runtime API" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.682745 4837 log.go:25] "Validated CRI v1 image API" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.684802 4837 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.691167 4837 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-01-06-56-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.691208 4837 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.707895 4837 manager.go:217] Machine: {Timestamp:2025-10-01 07:05:43.705009158 +0000 UTC m=+0.546616633 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1fa2a216-c629-41b0-a072-88b5c711fdce BootID:dd6f714d-42ba-4793-b8d3-f7eb0219041e Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a9:53:83 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a9:53:83 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:60:33:20 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1c:0b:73 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:cb:37:1f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:11:46:28 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:ce:d4:c3 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:5b:d8:65 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:84:a1:7a:6e:eb Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:17:08:3b:40:e1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.708177 4837 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.708340 4837 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.708667 4837 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.708954 4837 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.708995 4837 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.709228 4837 topology_manager.go:138] "Creating topology manager with none policy" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.709240 4837 container_manager_linux.go:303] "Creating device plugin manager" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.709792 4837 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.709960 4837 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.710797 4837 state_mem.go:36] "Initialized new in-memory state store" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.711627 4837 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.718419 4837 kubelet.go:418] "Attempting to sync node with API server" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.718497 4837 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.718553 4837 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.718581 4837 kubelet.go:324] "Adding apiserver pod source" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.718604 4837 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.724152 4837 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.724441 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.724461 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.724539 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.724551 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.725201 4837 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.727203 4837 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.728920 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.728994 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729018 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729041 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729073 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729093 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729117 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729147 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729171 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729192 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729218 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.729251 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.730325 4837 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.731388 4837 server.go:1280] "Started kubelet" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.731658 4837 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.732477 4837 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.732483 4837 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.733227 4837 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 01 07:05:43 crc systemd[1]: Started Kubernetes Kubelet. Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.735780 4837 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.735831 4837 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.736079 4837 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 02:39:17.973021418 +0000 UTC Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.736137 4837 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1747h33m34.236888821s for next certificate rotation Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.736961 4837 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.737144 4837 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.737211 4837 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.737884 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.77:6443: connect: connection refused" interval="200ms" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.738252 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.738495 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.738339 4837 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.738820 4837 factory.go:55] Registering systemd factory Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.738938 4837 factory.go:221] Registration of the systemd container factory successfully Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.736194 4837 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.738510 4837 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.77:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a4c29129e6812 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 07:05:43.73134133 +0000 UTC m=+0.572948835,LastTimestamp:2025-10-01 07:05:43.73134133 +0000 UTC m=+0.572948835,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.739774 4837 factory.go:153] Registering CRI-O factory Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.739985 4837 factory.go:221] Registration of the crio container factory successfully Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.740128 4837 factory.go:103] Registering Raw factory Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.740247 4837 manager.go:1196] Started watching for new ooms in manager Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.740443 4837 server.go:460] "Adding debug handlers to kubelet server" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.742384 4837 manager.go:319] Starting recovery of all containers Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.762778 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763293 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763421 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763542 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763676 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763861 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.763984 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764100 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764218 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764336 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764504 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764627 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764775 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.764914 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765031 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765167 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765290 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765409 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765539 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765665 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765827 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.765961 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766079 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766235 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766371 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766484 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766630 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766793 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.766912 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767023 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767132 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767281 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767397 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767513 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767623 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.767768 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768187 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768305 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768418 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768539 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768792 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.768929 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769061 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769179 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769290 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769402 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769516 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769642 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769791 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.769908 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770026 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770170 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770340 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770475 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770598 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770750 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770877 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.770997 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.771131 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.772435 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.772584 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.772736 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.772862 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.772997 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773114 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773227 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773350 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773468 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773598 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773746 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773874 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.773993 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774155 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774283 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774404 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774519 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774633 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774798 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.774931 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775065 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775197 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775312 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775440 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775556 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775677 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775830 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.775943 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.776077 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.778738 4837 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.778961 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779185 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779352 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779506 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779656 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779846 4837 manager.go:324] Recovery completed Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.779851 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.780195 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.780403 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.780587 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.780794 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.780993 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.781418 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.781565 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.781687 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.781837 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782020 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782202 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782348 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782475 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782598 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782813 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.782960 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783082 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783202 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783316 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783443 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783574 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783688 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783839 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.783968 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784102 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784238 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784396 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784538 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784664 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784843 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.784959 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785129 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785297 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785457 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785585 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785739 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.785900 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786016 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786157 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786307 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786441 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786577 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786755 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.786889 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787014 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787155 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787271 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787405 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787520 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787651 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787870 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.787990 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788122 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788237 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788378 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788516 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788666 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788854 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.788970 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789082 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789205 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789357 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789487 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789647 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789814 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.789945 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790064 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790238 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790387 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790469 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790544 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790622 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790709 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790795 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790876 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.790951 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791025 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791100 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791177 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791262 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791345 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791425 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791504 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791580 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791663 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791772 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791850 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.791927 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792005 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792116 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792203 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792282 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792397 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792485 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792574 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792662 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792805 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792903 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792707 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793020 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.792988 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793230 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793313 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793428 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793531 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793667 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793810 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.793930 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794029 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794111 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794325 4837 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794341 4837 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794364 4837 state_mem.go:36] "Initialized new in-memory state store" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794331 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794555 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794635 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794753 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794837 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.794937 4837 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.795022 4837 reconstruct.go:97] "Volume reconstruction finished" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.795105 4837 reconciler.go:26] "Reconciler: start to sync state" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.810248 4837 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.814537 4837 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.814613 4837 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.814664 4837 kubelet.go:2335] "Starting kubelet main sync loop" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.814819 4837 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 01 07:05:43 crc kubenswrapper[4837]: W1001 07:05:43.816407 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.816461 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.816637 4837 policy_none.go:49] "None policy: Start" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.819330 4837 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.819455 4837 state_mem.go:35] "Initializing new in-memory state store" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.840667 4837 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.891027 4837 manager.go:334] "Starting Device Plugin manager" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.891081 4837 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.891097 4837 server.go:79] "Starting device plugin registration server" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.891559 4837 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.891582 4837 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.892074 4837 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.892172 4837 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.892185 4837 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.901617 4837 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.915799 4837 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.915959 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.919498 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.919590 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.919721 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.920775 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.920954 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.921011 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922572 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922734 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922885 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.922997 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923001 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923032 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923116 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923534 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923555 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923863 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.923995 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.924075 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.924356 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.924440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.924455 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925301 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925343 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925356 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925572 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925598 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.925500 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.927596 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.927639 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.927651 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.927955 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.927990 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928452 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928655 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928875 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928916 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.928886 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.938779 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.77:6443: connect: connection refused" interval="400ms" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.992214 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.993517 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.993721 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.993885 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.994035 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:43 crc kubenswrapper[4837]: E1001 07:05:43.994748 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.77:6443: connect: connection refused" node="crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.996972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.997307 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.997473 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.997637 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.997841 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.997998 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998146 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998469 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998648 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998826 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.998991 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.999151 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.999285 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:43 crc kubenswrapper[4837]: I1001 07:05:43.999428 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.101172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.101504 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.101745 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.102023 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.102217 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.102425 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.102590 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.102605 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103028 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103267 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103506 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103622 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103105 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.103398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104447 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104447 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104535 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104566 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104594 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104623 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104666 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.104718 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105109 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105149 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105182 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.105273 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.195471 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.197974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.198061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.198087 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.198133 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.198930 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.77:6443: connect: connection refused" node="crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.264553 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.297353 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.320602 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.321389 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-dd82a33bd38b7e25f3a26fcee33b658a1a5e20816348c2df33c9f265fe053880 WatchSource:0}: Error finding container dd82a33bd38b7e25f3a26fcee33b658a1a5e20816348c2df33c9f265fe053880: Status 404 returned error can't find the container with id dd82a33bd38b7e25f3a26fcee33b658a1a5e20816348c2df33c9f265fe053880 Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.327433 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.334079 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.339749 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.77:6443: connect: connection refused" interval="800ms" Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.342353 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-8e407fc01756fad9fb385c286d9b0ff5b3ed1c3ce6f6bdbe6ed3df372a603bde WatchSource:0}: Error finding container 8e407fc01756fad9fb385c286d9b0ff5b3ed1c3ce6f6bdbe6ed3df372a603bde: Status 404 returned error can't find the container with id 8e407fc01756fad9fb385c286d9b0ff5b3ed1c3ce6f6bdbe6ed3df372a603bde Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.343549 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b28db19b9296961256e96beecf1b5e9e0f29e8867f1e87edf9b78c96b10f18bd WatchSource:0}: Error finding container b28db19b9296961256e96beecf1b5e9e0f29e8867f1e87edf9b78c96b10f18bd: Status 404 returned error can't find the container with id b28db19b9296961256e96beecf1b5e9e0f29e8867f1e87edf9b78c96b10f18bd Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.362711 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6a7f5464b98fdc713b0184f306eeea631be1b65fda0a91ff53574ddc0942505b WatchSource:0}: Error finding container 6a7f5464b98fdc713b0184f306eeea631be1b65fda0a91ff53574ddc0942505b: Status 404 returned error can't find the container with id 6a7f5464b98fdc713b0184f306eeea631be1b65fda0a91ff53574ddc0942505b Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.364597 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b8f2844d58ecf6fcf2fba1f30a922cf73f9c345f85aa3625769b70d721648a92 WatchSource:0}: Error finding container b8f2844d58ecf6fcf2fba1f30a922cf73f9c345f85aa3625769b70d721648a92: Status 404 returned error can't find the container with id b8f2844d58ecf6fcf2fba1f30a922cf73f9c345f85aa3625769b70d721648a92 Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.553160 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.553271 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.599742 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.601187 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.601240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.601255 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.601290 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.601850 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.77:6443: connect: connection refused" node="crc" Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.732716 4837 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.822220 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b8f2844d58ecf6fcf2fba1f30a922cf73f9c345f85aa3625769b70d721648a92"} Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.824515 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6a7f5464b98fdc713b0184f306eeea631be1b65fda0a91ff53574ddc0942505b"} Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.826318 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b28db19b9296961256e96beecf1b5e9e0f29e8867f1e87edf9b78c96b10f18bd"} Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.827599 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e407fc01756fad9fb385c286d9b0ff5b3ed1c3ce6f6bdbe6ed3df372a603bde"} Oct 01 07:05:44 crc kubenswrapper[4837]: I1001 07:05:44.829300 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dd82a33bd38b7e25f3a26fcee33b658a1a5e20816348c2df33c9f265fe053880"} Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.913588 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.913675 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:44 crc kubenswrapper[4837]: W1001 07:05:44.941180 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:44 crc kubenswrapper[4837]: E1001 07:05:44.941266 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:45 crc kubenswrapper[4837]: E1001 07:05:45.141485 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.77:6443: connect: connection refused" interval="1.6s" Oct 01 07:05:45 crc kubenswrapper[4837]: W1001 07:05:45.313931 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:45 crc kubenswrapper[4837]: E1001 07:05:45.314028 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.402858 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.405680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.405808 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.405829 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.405876 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:45 crc kubenswrapper[4837]: E1001 07:05:45.406522 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.77:6443: connect: connection refused" node="crc" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.733342 4837 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.834144 4837 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1" exitCode=0 Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.834235 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.834261 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.835673 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.835749 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.835766 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.838884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.838954 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.838977 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.840865 4837 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100" exitCode=0 Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.840959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.841102 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.842489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.842526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.842542 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.844287 4837 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8ba8caba2f576ce4a400ab260a4caa0ed5fdb0d0f104d878dcf223cd4330644a" exitCode=0 Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.844395 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8ba8caba2f576ce4a400ab260a4caa0ed5fdb0d0f104d878dcf223cd4330644a"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.844409 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.848678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.848750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.848765 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.848935 4837 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0" exitCode=0 Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.849010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0"} Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.849170 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.851158 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.851229 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.851243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.853336 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.854632 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.854671 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:45 crc kubenswrapper[4837]: I1001 07:05:45.854684 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:46 crc kubenswrapper[4837]: W1001 07:05:46.500996 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:46 crc kubenswrapper[4837]: E1001 07:05:46.501099 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.77:6443: connect: connection refused" logger="UnhandledError" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.733430 4837 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.77:6443: connect: connection refused Oct 01 07:05:46 crc kubenswrapper[4837]: E1001 07:05:46.743622 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.77:6443: connect: connection refused" interval="3.2s" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.856637 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.856704 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.856717 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.856726 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.860230 4837 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1bcfc3422464059102f5c29a0243dcba1f41b543bb7be70f63d733bc34514723" exitCode=0 Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.860319 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1bcfc3422464059102f5c29a0243dcba1f41b543bb7be70f63d733bc34514723"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.860425 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.866393 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.866454 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.866472 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.877023 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.877070 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.877089 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.877171 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.878451 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.878478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.878488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.881059 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"58bf5812f5a56f60cf83e3908a227ea3bf593692379d9371c27b51d03b93f017"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.881267 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.883666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.883732 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.883753 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.889593 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c"} Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.889757 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.890866 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.890900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:46 crc kubenswrapper[4837]: I1001 07:05:46.890917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.007335 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.009039 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.009075 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.009088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.009114 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:47 crc kubenswrapper[4837]: E1001 07:05:47.009614 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.77:6443: connect: connection refused" node="crc" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.265339 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.897237 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028"} Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.897280 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.898773 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.898823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.898847 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901088 4837 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="961223af01f2468c4df40c69233a5617c436f6515f8fc61dfeb7371bc11b0d80" exitCode=0 Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901253 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"961223af01f2468c4df40c69233a5617c436f6515f8fc61dfeb7371bc11b0d80"} Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901366 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901479 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901371 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.901617 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.902802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.902841 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.902858 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903713 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903768 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903787 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903872 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903903 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903921 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.903945 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.904000 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:47 crc kubenswrapper[4837]: I1001 07:05:47.904031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.911624 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.911710 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.911837 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.912435 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.912718 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"887f453ea44e43c9fed22087c0410b7c007db4ebb9df7bb1e20cc704c6d66c2e"} Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.912793 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"898ed42895ede50731534498dd15369aaf7f58633a4d922f9a8011193049daaf"} Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.912814 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4733d5e594097f9e5fafb55f72edd5d15cb99f8190adc870c433c17865cce5b3"} Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913424 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913438 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.913498 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.914118 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.914199 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:48 crc kubenswrapper[4837]: I1001 07:05:48.914220 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.922950 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"259d35bf285843ca56c65dc11fb9788392c018e5158e508096bcadefa253d3b0"} Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.923012 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5e5e6255ea5e6fc1fadfa20bf2e856b6e3f5b3530e1aae248d11ba50ec057af2"} Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.923126 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.925031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.925077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:49 crc kubenswrapper[4837]: I1001 07:05:49.925092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.210796 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.213231 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.213325 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.213346 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.213382 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.266201 4837 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.266312 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.266569 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.266775 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.266837 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.268344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.268392 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.268418 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.927180 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.928923 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.929010 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.929030 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.992766 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.992998 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.995039 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.995098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:50 crc kubenswrapper[4837]: I1001 07:05:50.995140 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.001167 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.503660 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.504081 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.506029 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.506087 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.506105 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.931597 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.933126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.933198 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:51 crc kubenswrapper[4837]: I1001 07:05:51.933217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.399326 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.399753 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.401982 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.402042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.402063 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.853737 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.854466 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.867468 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.867564 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:52 crc kubenswrapper[4837]: I1001 07:05:52.867594 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:53 crc kubenswrapper[4837]: I1001 07:05:53.322351 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:53 crc kubenswrapper[4837]: I1001 07:05:53.322794 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:53 crc kubenswrapper[4837]: I1001 07:05:53.325009 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:53 crc kubenswrapper[4837]: I1001 07:05:53.325081 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:53 crc kubenswrapper[4837]: I1001 07:05:53.325109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:53 crc kubenswrapper[4837]: E1001 07:05:53.901877 4837 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.154570 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.154873 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.157716 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.157780 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.157802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.161984 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.946115 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.948084 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.948165 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:55 crc kubenswrapper[4837]: I1001 07:05:55.948186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.734021 4837 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 01 07:05:57 crc kubenswrapper[4837]: W1001 07:05:57.826094 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.826241 4837 trace.go:236] Trace[1866698750]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 07:05:47.824) (total time: 10001ms): Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[1866698750]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:05:57.826) Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[1866698750]: [10.001433183s] [10.001433183s] END Oct 01 07:05:57 crc kubenswrapper[4837]: E1001 07:05:57.826286 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 01 07:05:57 crc kubenswrapper[4837]: W1001 07:05:57.914959 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.915112 4837 trace.go:236] Trace[74638808]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 07:05:47.913) (total time: 10001ms): Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[74638808]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:05:57.914) Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[74638808]: [10.001230416s] [10.001230416s] END Oct 01 07:05:57 crc kubenswrapper[4837]: E1001 07:05:57.915153 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.954126 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.956722 4837 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028" exitCode=255 Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.956773 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028"} Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.956958 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.958057 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.958093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.958108 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.958641 4837 scope.go:117] "RemoveContainer" containerID="09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028" Oct 01 07:05:57 crc kubenswrapper[4837]: W1001 07:05:57.991835 4837 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 01 07:05:57 crc kubenswrapper[4837]: I1001 07:05:57.991974 4837 trace.go:236] Trace[1678552378]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 07:05:47.989) (total time: 10001ms): Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[1678552378]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:05:57.991) Oct 01 07:05:57 crc kubenswrapper[4837]: Trace[1678552378]: [10.001935782s] [10.001935782s] END Oct 01 07:05:57 crc kubenswrapper[4837]: E1001 07:05:57.992008 4837 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.182272 4837 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.182351 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.186789 4837 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.186874 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.962396 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.964811 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539"} Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.965008 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.966344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.966394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:58 crc kubenswrapper[4837]: I1001 07:05:58.966412 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.557393 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.557679 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.559486 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.559552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.559564 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.593763 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.967591 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.969133 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.969211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.969237 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:05:59 crc kubenswrapper[4837]: I1001 07:05:59.986171 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.265360 4837 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.265485 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.289371 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.289596 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.289785 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.291056 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.291124 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.291152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.296517 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.971455 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.971502 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973368 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973384 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973405 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:00 crc kubenswrapper[4837]: I1001 07:06:00.973490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.726240 4837 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.748362 4837 apiserver.go:52] "Watching apiserver" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.754775 4837 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.756050 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.756563 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.756609 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:01 crc kubenswrapper[4837]: E1001 07:06:01.756655 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.756570 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:01 crc kubenswrapper[4837]: E1001 07:06:01.756746 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.756890 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.758150 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.759508 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.759809 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:01 crc kubenswrapper[4837]: E1001 07:06:01.759907 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765338 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765365 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765428 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765478 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765497 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.765888 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.766027 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.768617 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.812459 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.835223 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.838976 4837 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.851142 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.868584 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.889424 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.907746 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.924386 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.940516 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:01 crc kubenswrapper[4837]: I1001 07:06:01.996081 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 01 07:06:02 crc kubenswrapper[4837]: I1001 07:06:02.442623 4837 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.181099 4837 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.182022 4837 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.182133 4837 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.191534 4837 trace.go:236] Trace[877585764]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 07:05:50.089) (total time: 13101ms): Oct 01 07:06:03 crc kubenswrapper[4837]: Trace[877585764]: ---"Objects listed" error: 13101ms (07:06:03.191) Oct 01 07:06:03 crc kubenswrapper[4837]: Trace[877585764]: [13.101911559s] [13.101911559s] END Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.191577 4837 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281756 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281818 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281889 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281912 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281949 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281968 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.281988 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282006 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282049 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282072 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282104 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282121 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282204 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282220 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282238 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282281 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282308 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282384 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282405 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282438 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282453 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282468 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282485 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282647 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282678 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282723 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282773 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282809 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283122 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283199 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283221 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283239 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283270 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283321 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283360 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283380 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283401 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283449 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283470 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282236 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283506 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283528 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283551 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283667 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283763 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283817 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283833 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283868 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283887 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283902 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283918 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283951 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283967 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283985 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284040 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284067 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284086 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284130 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284149 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284166 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284199 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284216 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284234 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284296 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284321 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284345 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284459 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284480 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284496 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284549 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284654 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284721 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284745 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284761 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284806 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284831 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284887 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284917 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284939 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284987 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285011 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285078 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285105 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285154 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285178 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285670 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282266 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282502 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282496 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282517 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282764 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282856 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.282974 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283151 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283173 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283202 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283279 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283297 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283340 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283468 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283551 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283722 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283770 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283797 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283846 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283867 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.283971 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284017 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284047 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284105 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284115 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284139 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284333 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284524 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284582 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284615 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284659 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284675 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.286530 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.286557 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.287340 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284747 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284883 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284925 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284921 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.284944 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285056 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.285157 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289265 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289289 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289390 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289396 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289481 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289823 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.289847 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290069 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290163 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290215 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290326 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290392 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290598 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290634 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290678 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290809 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.290989 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.291146 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.291297 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.291486 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.292124 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.292167 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.292319 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.293172 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.293335 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.293351 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.293543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.293656 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294041 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294206 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294279 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294393 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294572 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294750 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.288878 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294855 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.294863 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295060 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295158 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295441 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295558 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295789 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295798 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295845 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295873 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295923 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295944 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295968 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.295989 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296010 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296030 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296051 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296072 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296093 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296117 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296222 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296255 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296284 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296321 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296342 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296364 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296389 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296410 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296431 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296452 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296473 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296517 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296538 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296559 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296581 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296601 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305244 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305360 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305432 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305511 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305581 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305668 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305856 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305944 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306038 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306106 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306264 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306338 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306404 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306481 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306569 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306662 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296281 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307256 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296482 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296960 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296991 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.297474 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.298131 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.298449 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.299376 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.299471 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.299503 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.299627 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.300046 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.301072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.301346 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.301377 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.300902 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.302770 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.303194 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.303506 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.304651 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305212 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305613 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305659 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305879 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.305994 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306058 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306110 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306256 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306409 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306377 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306657 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.306787 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:03.806755626 +0000 UTC m=+20.648363171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307724 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307764 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307791 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307819 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307840 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307885 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307909 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307931 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307957 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307984 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308006 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308028 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308082 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308114 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308165 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308200 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308232 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308261 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308291 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308318 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308334 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308350 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308382 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308414 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308446 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308476 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308498 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308519 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308543 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308563 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308585 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308607 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308627 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308649 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308672 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308720 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308417 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308749 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308776 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308823 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308846 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308872 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308895 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308924 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308970 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308998 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309032 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309064 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309093 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309128 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309159 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309229 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309280 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309321 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309387 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309417 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309458 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309493 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309525 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309557 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309591 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309628 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309662 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309717 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309817 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309840 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309859 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309877 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309895 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309912 4837 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309928 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309945 4837 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309960 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309979 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309998 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310017 4837 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310035 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310053 4837 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310074 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310090 4837 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310107 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310129 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310149 4837 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310166 4837 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310185 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310202 4837 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310221 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310237 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310253 4837 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310269 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310286 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310303 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310320 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310335 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310350 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310366 4837 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310383 4837 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310399 4837 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310414 4837 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310430 4837 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310445 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310461 4837 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310477 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310492 4837 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310507 4837 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310523 4837 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310540 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310559 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310579 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310596 4837 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310620 4837 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310637 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310655 4837 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310672 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310753 4837 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310781 4837 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310800 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310817 4837 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310835 4837 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310851 4837 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310868 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310884 4837 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310901 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310918 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310935 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310952 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310969 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310987 4837 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311005 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311020 4837 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311035 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311049 4837 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311065 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311081 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311098 4837 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311116 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311134 4837 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311150 4837 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311167 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311182 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311198 4837 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311214 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311235 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311251 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311266 4837 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311281 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311299 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311316 4837 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311330 4837 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311345 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.308725 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306935 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.312476 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.307090 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.296344 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309005 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309041 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309217 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.306807 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309209 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309551 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.309990 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310021 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310370 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.310558 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311071 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.311940 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.312024 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.312447 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.312642 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313048 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313207 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313287 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313286 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313744 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313932 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.313971 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.314192 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314310 4837 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314372 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314405 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314455 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314474 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314514 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314566 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314601 4837 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314616 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314629 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314644 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314681 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314682 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314726 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314745 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314796 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314813 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.314840 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:03.81481779 +0000 UTC m=+20.656425245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.314871 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.314926 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.314980 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:03.814971994 +0000 UTC m=+20.656579449 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.315003 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.315324 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.316066 4837 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.316077 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.316823 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.316988 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.317018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.317253 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.320157 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.321679 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.322768 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.326285 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.327338 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.331200 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.331239 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.331260 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.331346 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:03.831322149 +0000 UTC m=+20.672929704 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.334930 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.335622 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.335974 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.336024 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.336416 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.337214 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.337320 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.337435 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.337661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.337835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.338601 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.339503 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.342227 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.347547 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.347643 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.348252 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.349085 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.349216 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.349957 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.350328 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.351152 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.352799 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.353306 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.353598 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.353628 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.353646 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.353720 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:03.853679727 +0000 UTC m=+20.695287372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.353852 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.355868 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.355987 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356220 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356243 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356441 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356580 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356663 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.356992 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.357358 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.359333 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.364436 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.374892 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.391710 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416297 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416347 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416387 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416399 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416409 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416419 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416427 4837 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416435 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416443 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416452 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416461 4837 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416469 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416478 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416487 4837 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416496 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416504 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416512 4837 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416521 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416529 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416538 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416546 4837 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416556 4837 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416566 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416575 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416584 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416594 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416602 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416610 4837 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416619 4837 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416627 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416637 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416646 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416755 4837 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416765 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416773 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416781 4837 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416793 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416801 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416810 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416818 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416827 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416836 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416845 4837 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416853 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416862 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416870 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416879 4837 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416887 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416895 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416903 4837 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416911 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416919 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416927 4837 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416935 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416943 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416951 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416959 4837 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416967 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416981 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.416990 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417004 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417013 4837 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417020 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417028 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417036 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417044 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417052 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417061 4837 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417070 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417077 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417087 4837 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417100 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417108 4837 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417118 4837 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417126 4837 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417135 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417143 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417151 4837 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417159 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417166 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417174 4837 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417182 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417189 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417197 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417205 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417213 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417222 4837 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417230 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417237 4837 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417245 4837 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417253 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417260 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417267 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417275 4837 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417284 4837 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417291 4837 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417334 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.417887 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.587246 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.604227 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.617946 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 07:06:03 crc kubenswrapper[4837]: W1001 07:06:03.626242 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-fcb7fe74e577d7fda49624a3fbf2cda4b804e738a1ee7dd3f4e6fa44333ac7b5 WatchSource:0}: Error finding container fcb7fe74e577d7fda49624a3fbf2cda4b804e738a1ee7dd3f4e6fa44333ac7b5: Status 404 returned error can't find the container with id fcb7fe74e577d7fda49624a3fbf2cda4b804e738a1ee7dd3f4e6fa44333ac7b5 Oct 01 07:06:03 crc kubenswrapper[4837]: W1001 07:06:03.637915 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-a74aa91385362e624192f5379f9db6679ec6a20841fe2f607a331facd3a04034 WatchSource:0}: Error finding container a74aa91385362e624192f5379f9db6679ec6a20841fe2f607a331facd3a04034: Status 404 returned error can't find the container with id a74aa91385362e624192f5379f9db6679ec6a20841fe2f607a331facd3a04034 Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.706233 4837 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.815173 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.815300 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.815368 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.818314 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.820954 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.821053 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.823341 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.824114 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.824221 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.824281 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.824427 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.824496 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:04.824474399 +0000 UTC m=+21.666081874 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.824566 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:04.824555061 +0000 UTC m=+21.666162526 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.824619 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.824653 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:04.824643974 +0000 UTC m=+21.666251449 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.825022 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.826958 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.827957 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.829531 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.830363 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.832536 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.846446 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.851030 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.852248 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.853106 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.854875 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.857680 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.858473 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.860058 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.861361 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.864465 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.865259 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.865679 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.867466 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.868324 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.868862 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.869981 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.870406 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.871639 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.874117 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.875298 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.878327 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.878790 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.879465 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.880634 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.882555 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.883090 4837 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.883204 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.885557 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.886374 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.886813 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.888772 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.890280 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.891284 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.896067 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.896165 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.897076 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.898344 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.899205 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.901136 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.902604 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.904259 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.905241 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.906643 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.907678 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.908838 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.912206 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.912754 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.913810 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.914130 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.914966 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.915651 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.925864 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.926207 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926047 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926432 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926518 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926647 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:04.926624322 +0000 UTC m=+21.768231777 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926277 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926854 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.926922 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: E1001 07:06:03.927021 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:04.927009813 +0000 UTC m=+21.768617268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.931945 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.944128 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.956909 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.982527 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fcb7fe74e577d7fda49624a3fbf2cda4b804e738a1ee7dd3f4e6fa44333ac7b5"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.984753 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.984787 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.984800 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0e9841e0c12d7c68c93d6d3b2aff3c94a46286727c07f8a523ba09dabd9b01a3"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.987049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.987077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a74aa91385362e624192f5379f9db6679ec6a20841fe2f607a331facd3a04034"} Oct 01 07:06:03 crc kubenswrapper[4837]: I1001 07:06:03.999719 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.010650 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.023904 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.035987 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.049595 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.067212 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.083970 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.103810 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.114663 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.125903 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.145235 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.158769 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.171225 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.185831 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.834538 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.834678 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.834778 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.834788 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:06.834757058 +0000 UTC m=+23.676364503 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.834845 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:06.83482736 +0000 UTC m=+23.676434815 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.834871 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.835020 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.835063 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:06.835055416 +0000 UTC m=+23.676662871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.935481 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:04 crc kubenswrapper[4837]: I1001 07:06:04.935533 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935649 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935665 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935676 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935740 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:06.935725052 +0000 UTC m=+23.777332507 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935788 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935796 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935803 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:04 crc kubenswrapper[4837]: E1001 07:06:04.935822 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:06.935816214 +0000 UTC m=+23.777423669 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.114448 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tnhdf"] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.114872 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.121056 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.121269 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.122619 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.139482 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.150906 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.164035 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.176965 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.190039 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.211009 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.223964 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.235891 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.238391 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-hosts-file\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.238438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmx25\" (UniqueName: \"kubernetes.io/projected/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-kube-api-access-bmx25\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.339661 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-hosts-file\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.339727 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmx25\" (UniqueName: \"kubernetes.io/projected/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-kube-api-access-bmx25\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.340046 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-hosts-file\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.361155 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmx25\" (UniqueName: \"kubernetes.io/projected/71d89cf1-3c27-4aa7-ae3f-5e9284fbf555-kube-api-access-bmx25\") pod \"node-resolver-tnhdf\" (UID: \"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\") " pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.427918 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tnhdf" Oct 01 07:06:05 crc kubenswrapper[4837]: W1001 07:06:05.449049 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71d89cf1_3c27_4aa7_ae3f_5e9284fbf555.slice/crio-ee56b32f7f3469a3f655c122b83efa2642443d38bbde3e8cc6621d943ab4e16b WatchSource:0}: Error finding container ee56b32f7f3469a3f655c122b83efa2642443d38bbde3e8cc6621d943ab4e16b: Status 404 returned error can't find the container with id ee56b32f7f3469a3f655c122b83efa2642443d38bbde3e8cc6621d943ab4e16b Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.504065 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-lpk6v"] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.504543 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.508513 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.508573 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.508613 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.508995 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.509095 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.523608 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.537573 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.555666 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.575634 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.588180 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.601473 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.613100 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.627789 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.642049 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6e12bb16-7807-437f-927c-71dc8d0a27ec-rootfs\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.642107 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e12bb16-7807-437f-927c-71dc8d0a27ec-proxy-tls\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.642156 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e12bb16-7807-437f-927c-71dc8d0a27ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.642178 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m2hw\" (UniqueName: \"kubernetes.io/projected/6e12bb16-7807-437f-927c-71dc8d0a27ec-kube-api-access-5m2hw\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.642545 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.742570 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e12bb16-7807-437f-927c-71dc8d0a27ec-proxy-tls\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.742658 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e12bb16-7807-437f-927c-71dc8d0a27ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.742677 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m2hw\" (UniqueName: \"kubernetes.io/projected/6e12bb16-7807-437f-927c-71dc8d0a27ec-kube-api-access-5m2hw\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.742714 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6e12bb16-7807-437f-927c-71dc8d0a27ec-rootfs\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.742784 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6e12bb16-7807-437f-927c-71dc8d0a27ec-rootfs\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.743457 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e12bb16-7807-437f-927c-71dc8d0a27ec-mcd-auth-proxy-config\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.748428 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e12bb16-7807-437f-927c-71dc8d0a27ec-proxy-tls\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.759255 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m2hw\" (UniqueName: \"kubernetes.io/projected/6e12bb16-7807-437f-927c-71dc8d0a27ec-kube-api-access-5m2hw\") pod \"machine-config-daemon-lpk6v\" (UID: \"6e12bb16-7807-437f-927c-71dc8d0a27ec\") " pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.815935 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.815996 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.816067 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:05 crc kubenswrapper[4837]: E1001 07:06:05.816104 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:05 crc kubenswrapper[4837]: E1001 07:06:05.816256 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:05 crc kubenswrapper[4837]: E1001 07:06:05.816335 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.819522 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:06:05 crc kubenswrapper[4837]: W1001 07:06:05.832175 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e12bb16_7807_437f_927c_71dc8d0a27ec.slice/crio-72e1711950bcdce9612a06257c34944aaa9ce86f020e54791546dfddcf3df32e WatchSource:0}: Error finding container 72e1711950bcdce9612a06257c34944aaa9ce86f020e54791546dfddcf3df32e: Status 404 returned error can't find the container with id 72e1711950bcdce9612a06257c34944aaa9ce86f020e54791546dfddcf3df32e Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.861309 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-28t2p"] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.862085 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.864310 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.864476 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.864760 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.864884 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.865221 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.866514 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vf2cm"] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.867447 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-kztv6"] Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.867735 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.867896 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.872640 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.873135 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.873219 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.873841 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.873990 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.874220 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.874298 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.874495 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.874573 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.885776 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.905647 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.923813 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.939869 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945057 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945100 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cnibin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945120 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945138 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-multus-certs\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945161 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv8f8\" (UniqueName: \"kubernetes.io/projected/9533dad3-ca00-4500-bfcb-614a5ee75545-kube-api-access-zv8f8\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945185 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945204 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945227 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945286 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945304 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cni-binary-copy\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945343 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-cnibin\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945371 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945442 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-system-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945491 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945517 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-os-release\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945541 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-conf-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945576 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-daemon-config\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945622 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-binary-copy\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945643 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-tuning-conf-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945675 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945708 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945742 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-system-cni-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945762 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945777 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-hostroot\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945795 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-bin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945814 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-socket-dir-parent\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-kubelet\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945851 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-etc-kubernetes\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.945990 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpts\" (UniqueName: \"kubernetes.io/projected/7dedbee2-d2c0-49fb-ac76-aa7562c61211-kube-api-access-ddpts\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946055 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946099 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946118 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-k8s-cni-cncf-io\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946137 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-netns\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946200 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946219 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946294 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-multus\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946337 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946361 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5btcx\" (UniqueName: \"kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946403 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.946431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-os-release\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.957106 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.974062 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.986436 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:05Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.994633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038"} Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.997278 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc"} Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.997361 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"72e1711950bcdce9612a06257c34944aaa9ce86f020e54791546dfddcf3df32e"} Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.999777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tnhdf" event={"ID":"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555","Type":"ContainerStarted","Data":"7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0"} Oct 01 07:06:05 crc kubenswrapper[4837]: I1001 07:06:05.999807 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tnhdf" event={"ID":"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555","Type":"ContainerStarted","Data":"ee56b32f7f3469a3f655c122b83efa2642443d38bbde3e8cc6621d943ab4e16b"} Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.012450 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.031954 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.045079 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047464 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-tuning-conf-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047497 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047513 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047527 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047543 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-system-cni-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047557 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047574 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-hostroot\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047590 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-socket-dir-parent\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047599 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047604 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-bin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047666 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-kubelet\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047678 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-socket-dir-parent\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047683 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-etc-kubernetes\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047708 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047647 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-system-cni-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047718 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047751 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047761 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047751 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-kubelet\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047767 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-etc-kubernetes\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047718 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpts\" (UniqueName: \"kubernetes.io/projected/7dedbee2-d2c0-49fb-ac76-aa7562c61211-kube-api-access-ddpts\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047626 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-bin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047930 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-hostroot\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047947 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047974 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-k8s-cni-cncf-io\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.047992 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-netns\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048031 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048051 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048072 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5btcx\" (UniqueName: \"kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048107 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-multus\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048086 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-var-lib-cni-multus\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048160 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048046 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-k8s-cni-cncf-io\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048161 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048206 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048235 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-os-release\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048259 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048300 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048324 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cnibin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048346 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048367 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048390 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-multus-certs\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048416 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv8f8\" (UniqueName: \"kubernetes.io/projected/9533dad3-ca00-4500-bfcb-614a5ee75545-kube-api-access-zv8f8\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048427 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048443 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048472 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048495 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048520 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048553 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cni-binary-copy\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048578 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-cnibin\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048616 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-system-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048653 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048670 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048693 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048676 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-netns\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048731 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cnibin\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048763 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-os-release\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048788 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048817 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048856 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048859 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048868 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048656 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-os-release\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048826 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-conf-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048892 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-host-run-multus-certs\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-conf-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048961 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-os-release\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048773 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048969 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-daemon-config\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049019 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-binary-copy\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.048678 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049239 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7dedbee2-d2c0-49fb-ac76-aa7562c61211-system-cni-dir\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049332 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-cni-binary-copy\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049326 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-cnibin\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9533dad3-ca00-4500-bfcb-614a5ee75545-tuning-conf-dir\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.049617 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.050003 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.050186 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7dedbee2-d2c0-49fb-ac76-aa7562c61211-multus-daemon-config\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.050241 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-binary-copy\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.052156 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9533dad3-ca00-4500-bfcb-614a5ee75545-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.052833 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.074754 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpts\" (UniqueName: \"kubernetes.io/projected/7dedbee2-d2c0-49fb-ac76-aa7562c61211-kube-api-access-ddpts\") pod \"multus-kztv6\" (UID: \"7dedbee2-d2c0-49fb-ac76-aa7562c61211\") " pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.075373 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv8f8\" (UniqueName: \"kubernetes.io/projected/9533dad3-ca00-4500-bfcb-614a5ee75545-kube-api-access-zv8f8\") pod \"multus-additional-cni-plugins-28t2p\" (UID: \"9533dad3-ca00-4500-bfcb-614a5ee75545\") " pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.076682 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5btcx\" (UniqueName: \"kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx\") pod \"ovnkube-node-vf2cm\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.079177 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.098212 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.110723 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.125380 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.139392 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.161006 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.177132 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.183263 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-28t2p" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.194349 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.205857 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.210632 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.212727 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-kztv6" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.227925 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: W1001 07:06:06.247795 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod401d9e9e_cba5_413c_b078_83858883db16.slice/crio-99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee WatchSource:0}: Error finding container 99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee: Status 404 returned error can't find the container with id 99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.252231 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.274495 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:06Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.859748 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.859928 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:10.859903032 +0000 UTC m=+27.701510487 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.860000 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.860038 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.860168 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.860223 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:10.86021224 +0000 UTC m=+27.701819695 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.860253 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.860368 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:10.860337814 +0000 UTC m=+27.701945269 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.960976 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:06 crc kubenswrapper[4837]: I1001 07:06:06.961044 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961196 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961213 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961226 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961269 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961317 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961332 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961285 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:10.961270936 +0000 UTC m=+27.802878391 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:06 crc kubenswrapper[4837]: E1001 07:06:06.961440 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:10.961416129 +0000 UTC m=+27.803023584 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.005022 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.007412 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerStarted","Data":"5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.007440 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerStarted","Data":"e2cc85d9bb44f6562da040d75eb34405b3d7d726d42789c2e00a921e26edcf1a"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.009598 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a" exitCode=0 Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.009679 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.009777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.013236 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce" exitCode=0 Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.013310 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.013347 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerStarted","Data":"719a5d5f2996ad818fbe54488121dc6a05c4ddea051e22fb15970685784d69c9"} Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.027510 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.027595 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zdszv"] Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.028599 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.035055 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.035399 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.035606 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.035759 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.053938 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.072546 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.091835 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.112553 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.131418 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.159972 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.163199 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn4ks\" (UniqueName: \"kubernetes.io/projected/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-kube-api-access-sn4ks\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.163285 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-serviceca\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.163368 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-host\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.175310 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.192567 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.208412 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.224471 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.241921 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.257029 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.264365 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn4ks\" (UniqueName: \"kubernetes.io/projected/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-kube-api-access-sn4ks\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.264421 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-serviceca\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.264461 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-host\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.264520 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-host\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.265357 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-serviceca\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.273146 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.279932 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.280963 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.282218 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.288826 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn4ks\" (UniqueName: \"kubernetes.io/projected/e5a30c54-f521-4faf-9a65-0e85b3ebe8c0-kube-api-access-sn4ks\") pod \"node-ca-zdszv\" (UID: \"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\") " pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.297174 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.323340 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.342856 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.355917 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.368302 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.385589 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.402968 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.403746 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zdszv" Oct 01 07:06:07 crc kubenswrapper[4837]: W1001 07:06:07.427476 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a30c54_f521_4faf_9a65_0e85b3ebe8c0.slice/crio-2074d9ce2c6d3193fd28a74eda1b476f23b61252a76f7e50aad66008240a31e8 WatchSource:0}: Error finding container 2074d9ce2c6d3193fd28a74eda1b476f23b61252a76f7e50aad66008240a31e8: Status 404 returned error can't find the container with id 2074d9ce2c6d3193fd28a74eda1b476f23b61252a76f7e50aad66008240a31e8 Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.429682 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.456751 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.475929 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.492415 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.510103 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.530088 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.547109 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.563959 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.578133 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.596143 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.609082 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.628499 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.642024 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.661822 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.675133 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.691053 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.706617 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.718314 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:07Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.815676 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.815783 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:07 crc kubenswrapper[4837]: E1001 07:06:07.815912 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:07 crc kubenswrapper[4837]: E1001 07:06:07.816000 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:07 crc kubenswrapper[4837]: I1001 07:06:07.816119 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:07 crc kubenswrapper[4837]: E1001 07:06:07.816184 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.020561 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerStarted","Data":"3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf"} Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.023954 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c"} Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.024010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31"} Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.025102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zdszv" event={"ID":"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0","Type":"ContainerStarted","Data":"2074d9ce2c6d3193fd28a74eda1b476f23b61252a76f7e50aad66008240a31e8"} Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.040401 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.058632 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.072538 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.088266 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.106810 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.125058 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.143462 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.166964 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.183994 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.205243 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.222060 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.239144 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.258553 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:08 crc kubenswrapper[4837]: I1001 07:06:08.287059 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:08Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.030936 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zdszv" event={"ID":"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0","Type":"ContainerStarted","Data":"04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.032938 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf" exitCode=0 Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.033007 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.039091 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.039135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.039147 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.039157 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.055584 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.075411 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.105313 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.124361 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.141364 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.154323 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.168253 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.181627 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.195263 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.207869 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.223867 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.237181 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.252923 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.274661 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.297735 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.312148 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.335222 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.353483 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.371945 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.388313 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.405962 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.425995 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.442808 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.461102 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.480012 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.496992 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.518303 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.537125 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.583222 4837 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.587480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.587774 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.587806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.587980 4837 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.604030 4837 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.604297 4837 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.605693 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.605757 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.605772 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.605793 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.605808 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.627464 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.633188 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.633336 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.633416 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.633511 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.633606 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.650897 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.657077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.657107 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.657532 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.657592 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.657603 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.677469 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.681327 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.681396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.681418 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.681445 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.681465 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.698268 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.702935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.703008 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.703062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.703096 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.703122 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.719359 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:09Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.719621 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.721608 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.721650 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.721664 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.721683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.721721 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.815331 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.815426 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.815331 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.815606 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.815778 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:09 crc kubenswrapper[4837]: E1001 07:06:09.816111 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.824872 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.824974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.824995 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.825065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.825086 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.929565 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.929628 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.929642 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.929663 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:09 crc kubenswrapper[4837]: I1001 07:06:09.929677 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:09Z","lastTransitionTime":"2025-10-01T07:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.033061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.033520 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.033677 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.033945 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.034131 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.047954 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61" exitCode=0 Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.048073 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.073431 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.098175 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.120197 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.137779 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.137834 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.137851 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.137876 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.137894 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.145409 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.168465 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.189776 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.208275 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.237883 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.240939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.240985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.240998 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.241023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.241040 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.256282 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.273698 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.292247 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.312294 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.328796 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.343655 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.343751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.343776 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.343806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.343825 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.345275 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.448783 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.448839 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.448859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.448886 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.448904 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.552673 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.552755 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.552770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.552794 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.552808 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.656098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.656164 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.656179 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.656203 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.656219 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.759645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.759756 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.759776 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.759805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.759827 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.863052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.863128 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.863146 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.863171 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.863186 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.905773 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:10 crc kubenswrapper[4837]: E1001 07:06:10.906085 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:18.906040084 +0000 UTC m=+35.747647579 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.906141 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.906198 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:10 crc kubenswrapper[4837]: E1001 07:06:10.906378 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:10 crc kubenswrapper[4837]: E1001 07:06:10.906389 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:10 crc kubenswrapper[4837]: E1001 07:06:10.906466 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:18.906449585 +0000 UTC m=+35.748057070 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:10 crc kubenswrapper[4837]: E1001 07:06:10.906493 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:18.906481685 +0000 UTC m=+35.748089180 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.965778 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.965826 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.965835 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.965853 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:10 crc kubenswrapper[4837]: I1001 07:06:10.965864 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:10Z","lastTransitionTime":"2025-10-01T07:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.007884 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.007966 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008159 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008197 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008221 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008233 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008238 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008254 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008317 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:19.00829429 +0000 UTC m=+35.849901755 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.008340 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:19.008331121 +0000 UTC m=+35.849938586 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.059276 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.062676 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1" exitCode=0 Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.062759 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.068899 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.068962 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.068980 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.069006 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.069023 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.090117 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.108247 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.137410 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.155391 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.173392 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.174032 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.174101 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.174121 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.174655 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.174769 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.195487 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.219319 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.239525 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.256650 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.277294 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.279269 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.279298 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.279308 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.279325 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.279337 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.296932 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.313946 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.330476 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.344232 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.382223 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.382291 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.382308 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.382332 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.382348 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.487562 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.487619 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.487638 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.487663 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.487683 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.509453 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.522890 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.541688 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.563018 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.575375 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.590337 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.590392 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.590411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.590435 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.590454 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.607406 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.622519 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.637170 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.656409 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.673112 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.690235 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.694195 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.694290 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.694318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.694357 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.694385 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.707626 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.722589 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.736404 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.751883 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:11Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.798017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.798410 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.798475 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.798544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.798611 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.815521 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.815672 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.815560 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.815741 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.815950 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:11 crc kubenswrapper[4837]: E1001 07:06:11.816102 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.902816 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.902897 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.902917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.902951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:11 crc kubenswrapper[4837]: I1001 07:06:11.902973 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:11Z","lastTransitionTime":"2025-10-01T07:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.005793 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.005875 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.005900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.005933 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.005958 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.070609 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerStarted","Data":"89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.087168 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.102951 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.109367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.109460 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.109484 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.109514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.109533 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.120772 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.138316 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.168622 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.199057 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.212503 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.212566 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.212583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.212608 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.212627 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.223116 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.241420 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.255338 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.267591 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.283331 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.294882 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.315715 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.315804 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.315825 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.315854 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.315873 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.318828 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.339928 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:12Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.419225 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.419661 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.419750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.419884 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.419952 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.523767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.524293 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.524485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.524675 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.524881 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.628078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.628150 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.628162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.628184 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.628196 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.731971 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.732025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.732042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.732066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.732082 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.836341 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.836920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.836935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.836957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.836973 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.939854 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.939922 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.939941 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.939968 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:12 crc kubenswrapper[4837]: I1001 07:06:12.939987 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:12Z","lastTransitionTime":"2025-10-01T07:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.043172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.043245 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.043271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.043306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.043332 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.078857 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb" exitCode=0 Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.078964 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.085575 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.086359 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.101061 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.116502 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.119450 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.143946 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.148882 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.148934 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.148954 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.148982 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.148999 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.160156 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.174352 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.192291 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.208135 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.223655 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.238301 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.252120 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.252176 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.252193 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.252221 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.252242 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.254055 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.275750 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.288454 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.304090 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.322185 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.339538 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.351220 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.355102 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.355146 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.355162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.355184 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.355198 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.365838 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.381015 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.393271 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.412290 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.428006 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.458496 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.458541 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.458552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.458587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.458599 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.462146 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.484423 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.500264 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.519111 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.539562 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.554997 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.561869 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.561907 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.561921 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.561941 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.561954 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.575158 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.665111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.665173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.665191 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.665214 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.665229 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.769568 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.769618 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.769631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.769669 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.769684 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.815457 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.815491 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.815457 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:13 crc kubenswrapper[4837]: E1001 07:06:13.815601 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:13 crc kubenswrapper[4837]: E1001 07:06:13.815776 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:13 crc kubenswrapper[4837]: E1001 07:06:13.815910 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.835631 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.854122 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.873066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.873123 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.873144 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.873178 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.873200 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.881761 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.902058 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.926153 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.944866 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.966359 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.976073 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.976194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.976215 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.976246 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.976267 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:13Z","lastTransitionTime":"2025-10-01T07:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:13 crc kubenswrapper[4837]: I1001 07:06:13.986759 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.005873 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.022385 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.039083 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.055352 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.071361 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.079643 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.079879 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.080374 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.080475 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.080523 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.090418 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.095005 4837 generic.go:334] "Generic (PLEG): container finished" podID="9533dad3-ca00-4500-bfcb-614a5ee75545" containerID="db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec" exitCode=0 Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.095084 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerDied","Data":"db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.095249 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.095825 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.114852 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.130298 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.131716 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.148737 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.166556 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.179284 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.184730 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.184770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.184784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.184804 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.184817 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.192424 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.206811 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.218975 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.230030 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.242439 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.256487 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.267401 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.285922 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.289977 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.290011 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.290024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.290043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.290054 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.300159 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.312620 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.325865 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.340571 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.365379 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.385442 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.392980 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.393026 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.393037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.393055 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.393069 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.417484 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.441636 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.457007 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.475767 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.496114 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.496165 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.496178 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.496201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.496216 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.498332 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.517352 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.531854 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.550108 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.563609 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.599082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.599154 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.599173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.599205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.599230 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.702629 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.702676 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.702688 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.702720 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.702731 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.806277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.806389 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.806410 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.806436 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.806454 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.913264 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.913335 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.913352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.913377 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:14 crc kubenswrapper[4837]: I1001 07:06:14.913401 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:14Z","lastTransitionTime":"2025-10-01T07:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.017830 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.017901 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.017919 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.017946 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.017965 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.106004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" event={"ID":"9533dad3-ca00-4500-bfcb-614a5ee75545","Type":"ContainerStarted","Data":"4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.106125 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.120989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.121048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.121062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.121086 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.121107 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.138158 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.161099 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.188483 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.209766 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.224426 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.224475 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.224495 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.224518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.224530 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.230389 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.257406 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.272784 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.285640 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.296946 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.309342 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.320118 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.327355 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.327407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.327418 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.327433 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.327443 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.331237 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.344122 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.361733 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:15Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.430166 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.430217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.430226 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.430238 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.430263 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.533796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.533845 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.533855 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.533874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.533888 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.638439 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.638491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.638506 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.638530 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.638545 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.741933 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.741981 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.741990 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.742004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.742013 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.815981 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:15 crc kubenswrapper[4837]: E1001 07:06:15.816112 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.815994 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:15 crc kubenswrapper[4837]: E1001 07:06:15.816244 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.816255 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:15 crc kubenswrapper[4837]: E1001 07:06:15.816635 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.845456 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.845525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.845542 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.845572 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.845591 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.954176 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.954236 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.954247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.954267 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:15 crc kubenswrapper[4837]: I1001 07:06:15.954279 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:15Z","lastTransitionTime":"2025-10-01T07:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.057457 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.057528 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.057550 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.057580 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.057596 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.114435 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/0.log" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.120298 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900" exitCode=1 Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.120402 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.121977 4837 scope.go:117] "RemoveContainer" containerID="3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.150235 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.161157 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.161190 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.161199 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.161214 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.161224 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.165014 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.191160 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.211838 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.227359 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.249811 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.265743 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.265998 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.266831 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.266929 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.267035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.267126 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.281010 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.295527 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.312025 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.329665 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.344553 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.368198 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"message\\\":\\\"al\\\\nI1001 07:06:15.837102 6095 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:06:15.837123 6095 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:06:15.837148 6095 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 07:06:15.837180 6095 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 07:06:15.838006 6095 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838257 6095 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:06:15.838277 6095 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838336 6095 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:06:15.838350 6095 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1001 07:06:15.839123 6095 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.370478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.370513 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.370524 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.370546 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.370562 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.386033 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:16Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.473752 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.473802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.473814 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.473833 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.473849 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.576556 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.576631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.576652 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.576680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.576732 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.679410 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.679513 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.679532 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.679560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.679583 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.782585 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.782650 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.782669 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.782723 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.782746 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.886789 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.886830 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.886841 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.886857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.886872 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.989725 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.989756 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.989764 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.989778 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:16 crc kubenswrapper[4837]: I1001 07:06:16.989787 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:16Z","lastTransitionTime":"2025-10-01T07:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.091639 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.091670 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.091679 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.091719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.091728 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.124954 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/0.log" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.128044 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.128148 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.143444 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.173134 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"message\\\":\\\"al\\\\nI1001 07:06:15.837102 6095 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:06:15.837123 6095 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:06:15.837148 6095 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 07:06:15.837180 6095 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 07:06:15.838006 6095 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838257 6095 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:06:15.838277 6095 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838336 6095 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:06:15.838350 6095 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1001 07:06:15.839123 6095 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.194317 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.195551 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.195584 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.195593 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.195611 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.195622 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.211155 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.227478 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.251585 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.270395 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.285429 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.298610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.298662 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.298676 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.298719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.298736 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.320203 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.349639 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.378108 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.392199 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.400873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.400907 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.400917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.400932 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.400942 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.406156 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.420382 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.503773 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.503872 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.503891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.503920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.503937 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.527405 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.607437 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.607490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.607503 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.607526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.607545 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.710946 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.711023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.711052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.711088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.711116 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.814417 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.814492 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.814503 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.814525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.814537 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.815033 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.815075 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:17 crc kubenswrapper[4837]: E1001 07:06:17.815273 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.815654 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:17 crc kubenswrapper[4837]: E1001 07:06:17.815814 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:17 crc kubenswrapper[4837]: E1001 07:06:17.816134 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.918440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.918498 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.918509 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.918528 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:17 crc kubenswrapper[4837]: I1001 07:06:17.918542 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:17Z","lastTransitionTime":"2025-10-01T07:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.021634 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.021670 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.021682 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.021724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.021740 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.124610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.124660 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.124671 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.124713 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.124727 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.134481 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/1.log" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.135552 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/0.log" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.139592 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001" exitCode=1 Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.139777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.139962 4837 scope.go:117] "RemoveContainer" containerID="3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.140956 4837 scope.go:117] "RemoveContainer" containerID="6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001" Oct 01 07:06:18 crc kubenswrapper[4837]: E1001 07:06:18.141262 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.181038 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"message\\\":\\\"al\\\\nI1001 07:06:15.837102 6095 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:06:15.837123 6095 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:06:15.837148 6095 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 07:06:15.837180 6095 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 07:06:15.838006 6095 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838257 6095 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:06:15.838277 6095 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838336 6095 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:06:15.838350 6095 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1001 07:06:15.839123 6095 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.206527 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.227794 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.227868 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.227888 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.227916 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.227938 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.228215 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.243580 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.272100 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.294119 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.309304 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.328626 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.331180 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.331383 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.331811 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.332097 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.332291 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.350837 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.369900 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.387195 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.409679 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.417405 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb"] Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.418103 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.420371 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.421428 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.430392 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.435606 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.435657 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.435749 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.435770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.435808 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.450840 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.467886 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.490051 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"message\\\":\\\"al\\\\nI1001 07:06:15.837102 6095 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:06:15.837123 6095 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:06:15.837148 6095 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 07:06:15.837180 6095 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 07:06:15.838006 6095 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838257 6095 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:06:15.838277 6095 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838336 6095 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:06:15.838350 6095 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1001 07:06:15.839123 6095 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.496724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.496794 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f47sj\" (UniqueName: \"kubernetes.io/projected/09e0351f-93e5-44bc-803c-5a722722f722-kube-api-access-f47sj\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.496828 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.496861 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09e0351f-93e5-44bc-803c-5a722722f722-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.506399 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.518656 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.533486 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.539078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.539130 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.539139 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.539156 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.539169 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.558508 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.576629 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.586170 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.598339 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f47sj\" (UniqueName: \"kubernetes.io/projected/09e0351f-93e5-44bc-803c-5a722722f722-kube-api-access-f47sj\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.598394 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.598416 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09e0351f-93e5-44bc-803c-5a722722f722-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.598468 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.599268 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.599556 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09e0351f-93e5-44bc-803c-5a722722f722-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.602895 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.607294 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09e0351f-93e5-44bc-803c-5a722722f722-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.629477 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.630502 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f47sj\" (UniqueName: \"kubernetes.io/projected/09e0351f-93e5-44bc-803c-5a722722f722-kube-api-access-f47sj\") pod \"ovnkube-control-plane-749d76644c-s9qpb\" (UID: \"09e0351f-93e5-44bc-803c-5a722722f722\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.642136 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.642209 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.642219 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.642232 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.642242 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.651270 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.664060 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.678002 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.691281 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.707909 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:18Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.732416 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.744263 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.744346 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.744378 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.744415 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.744442 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: W1001 07:06:18.747844 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09e0351f_93e5_44bc_803c_5a722722f722.slice/crio-cd6c4b3939ba1f9f99dd636a0f47a493c9e9f5af26000e492cb0975c1ecabf06 WatchSource:0}: Error finding container cd6c4b3939ba1f9f99dd636a0f47a493c9e9f5af26000e492cb0975c1ecabf06: Status 404 returned error can't find the container with id cd6c4b3939ba1f9f99dd636a0f47a493c9e9f5af26000e492cb0975c1ecabf06 Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.854016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.854069 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.854084 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.854103 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.854117 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.958096 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.958180 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.958204 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.958235 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:18 crc kubenswrapper[4837]: I1001 07:06:18.958268 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:18Z","lastTransitionTime":"2025-10-01T07:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.003137 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.003320 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:06:35.003295855 +0000 UTC m=+51.844903330 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.003400 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.003449 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.003544 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.003601 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:35.003589483 +0000 UTC m=+51.845196938 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.003677 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.003830 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:35.003799318 +0000 UTC m=+51.845406763 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.061260 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.061315 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.061332 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.061356 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.061373 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.104893 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.104943 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105132 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105154 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105167 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105163 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105200 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105217 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105224 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:35.105208323 +0000 UTC m=+51.946815778 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.105291 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:06:35.105261474 +0000 UTC m=+51.946868939 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.146819 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" event={"ID":"09e0351f-93e5-44bc-803c-5a722722f722","Type":"ContainerStarted","Data":"fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.146892 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" event={"ID":"09e0351f-93e5-44bc-803c-5a722722f722","Type":"ContainerStarted","Data":"0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.146903 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" event={"ID":"09e0351f-93e5-44bc-803c-5a722722f722","Type":"ContainerStarted","Data":"cd6c4b3939ba1f9f99dd636a0f47a493c9e9f5af26000e492cb0975c1ecabf06"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.149251 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-z66wh"] Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.149914 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.149997 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.150048 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/1.log" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.162595 4837 scope.go:117] "RemoveContainer" containerID="6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.162844 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.163594 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.163635 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.163650 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.163671 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.163686 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.170949 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6a17c312d3a13bbc72f8d6f27e67cea110ba1a3cd94297643f478a5467f900\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"message\\\":\\\"al\\\\nI1001 07:06:15.837102 6095 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:06:15.837123 6095 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:06:15.837148 6095 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1001 07:06:15.837180 6095 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 07:06:15.838006 6095 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838257 6095 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:06:15.838277 6095 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1001 07:06:15.838336 6095 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:06:15.838350 6095 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1001 07:06:15.839123 6095 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.188201 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.203314 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.205752 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.205801 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfs6f\" (UniqueName: \"kubernetes.io/projected/95803be9-8a27-4569-958d-81666ad8defc-kube-api-access-vfs6f\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.220541 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.244047 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.260531 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.266084 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.266122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.266132 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.266149 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.266159 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.274210 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.290600 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.304820 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.307479 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.307541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfs6f\" (UniqueName: \"kubernetes.io/projected/95803be9-8a27-4569-958d-81666ad8defc-kube-api-access-vfs6f\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.307647 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.307741 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:19.807721394 +0000 UTC m=+36.649328849 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.318432 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.325498 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfs6f\" (UniqueName: \"kubernetes.io/projected/95803be9-8a27-4569-958d-81666ad8defc-kube-api-access-vfs6f\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.330800 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.341300 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.356304 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.369065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.369098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.369110 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.369158 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.369172 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.371372 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.387315 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.410345 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.425519 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.440671 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.453001 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.472977 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.473037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.473053 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.473073 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.473086 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.473087 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.488683 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.501673 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.515505 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.530128 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.543939 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.557505 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.573873 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.576332 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.576435 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.576454 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.576482 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.576500 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.590253 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.604050 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.623446 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.638586 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:19Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.680092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.680158 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.680174 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.680198 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.680211 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.783375 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.783444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.783462 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.783488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.783511 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.814323 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.814502 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.814575 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:20.814556512 +0000 UTC m=+37.656163967 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.815022 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.815068 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.815033 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.815196 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.815297 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:19 crc kubenswrapper[4837]: E1001 07:06:19.815465 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.887174 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.887243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.887254 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.887268 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.887277 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.989564 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.989606 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.989615 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.989630 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:19 crc kubenswrapper[4837]: I1001 07:06:19.989640 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:19Z","lastTransitionTime":"2025-10-01T07:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.000994 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.001034 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.001044 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.001062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.001123 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.021391 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:20Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.025969 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.026016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.026028 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.026046 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.026058 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.040777 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:20Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.044786 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.044839 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.044853 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.044873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.044886 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.058743 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:20Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.062275 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.062309 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.062321 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.062510 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.062555 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.081914 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:20Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.086229 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.086292 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.086306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.086323 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.086336 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.102794 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:20Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.103063 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.105106 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.105186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.105212 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.105245 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.105269 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.208893 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.208977 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.209002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.209040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.209066 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.312599 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.312650 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.312667 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.312719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.312741 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.416636 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.416680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.416705 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.416725 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.416739 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.521144 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.521233 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.521281 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.521317 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.521343 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.624983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.625093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.625118 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.625152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.625174 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.728016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.728092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.728105 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.728126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.728139 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.815568 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.815809 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.824646 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.824882 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:20 crc kubenswrapper[4837]: E1001 07:06:20.825144 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:22.825115467 +0000 UTC m=+39.666722972 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.831516 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.831554 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.831618 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.831649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.831667 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.935069 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.935145 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.935162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.935185 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:20 crc kubenswrapper[4837]: I1001 07:06:20.935203 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:20Z","lastTransitionTime":"2025-10-01T07:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.039061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.039103 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.039121 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.039146 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.039163 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.144617 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.144657 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.144666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.144684 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.144721 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.247943 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.248062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.248120 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.248153 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.248178 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.351336 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.351397 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.351415 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.351478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.351497 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.454874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.454951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.454974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.455007 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.455030 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.558310 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.558372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.558388 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.558412 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.558430 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.660836 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.660922 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.660946 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.660975 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.660994 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.765004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.765086 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.765110 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.765140 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.765160 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.815994 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.816059 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.816017 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:21 crc kubenswrapper[4837]: E1001 07:06:21.816223 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:21 crc kubenswrapper[4837]: E1001 07:06:21.816386 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:21 crc kubenswrapper[4837]: E1001 07:06:21.816563 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.870368 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.870462 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.870496 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.870573 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.870607 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.975425 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.975512 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.975547 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.975584 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:21 crc kubenswrapper[4837]: I1001 07:06:21.975610 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:21Z","lastTransitionTime":"2025-10-01T07:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.079302 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.079394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.079408 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.079449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.079468 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.181987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.182043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.182055 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.182080 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.182093 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.284828 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.284873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.284881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.284898 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.284907 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.388352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.388425 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.388443 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.388471 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.388490 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.494025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.494081 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.494093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.494114 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.494128 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.597331 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.597372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.597385 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.597406 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.597421 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.699921 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.700002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.700024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.700059 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.700082 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.803900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.803978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.803997 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.804029 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.804047 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.815269 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:22 crc kubenswrapper[4837]: E1001 07:06:22.815480 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.854198 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:22 crc kubenswrapper[4837]: E1001 07:06:22.854431 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:22 crc kubenswrapper[4837]: E1001 07:06:22.854582 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:26.85454077 +0000 UTC m=+43.696148305 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.908188 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.908243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.908259 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.908287 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:22 crc kubenswrapper[4837]: I1001 07:06:22.908304 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:22Z","lastTransitionTime":"2025-10-01T07:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.011956 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.012004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.012018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.012036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.012048 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.116910 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.116968 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.116986 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.117012 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.117031 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.220278 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.220362 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.220400 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.220436 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.220457 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.324120 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.324172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.324184 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.324208 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.324221 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.428104 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.428173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.428192 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.428222 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.428240 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.532652 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.532775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.532801 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.532835 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.532858 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.637021 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.637065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.637079 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.637102 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.637116 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.741242 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.741319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.741338 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.741366 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.741389 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.815461 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:23 crc kubenswrapper[4837]: E1001 07:06:23.815677 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.815795 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.816084 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:23 crc kubenswrapper[4837]: E1001 07:06:23.817539 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:23 crc kubenswrapper[4837]: E1001 07:06:23.817728 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.842952 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.844528 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.844596 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.844624 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.844659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.844686 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.867538 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.892262 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.909365 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.938567 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.947488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.947550 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.947569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.947593 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.947610 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:23Z","lastTransitionTime":"2025-10-01T07:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.959486 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.978593 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:23 crc kubenswrapper[4837]: I1001 07:06:23.994453 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.019933 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.040933 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.053532 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.053607 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.053620 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.053662 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.053675 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.055821 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.072527 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.084845 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.096789 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.110471 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.120272 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.157007 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.157054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.157068 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.157086 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.157098 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.260380 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.260445 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.260463 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.260489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.260507 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.363138 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.363169 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.363177 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.363190 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.363198 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.466951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.467023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.467047 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.467077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.467099 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.570061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.570175 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.570203 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.570275 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.570345 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.672895 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.672952 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.672965 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.672985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.672997 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.779027 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.779079 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.779092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.779109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.779120 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.815347 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:24 crc kubenswrapper[4837]: E1001 07:06:24.815540 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.881806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.881846 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.881859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.881877 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.881889 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.984667 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.984755 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.984767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.984781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:24 crc kubenswrapper[4837]: I1001 07:06:24.984790 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:24Z","lastTransitionTime":"2025-10-01T07:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.086681 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.086781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.086797 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.086822 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.086842 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.189569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.189631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.189649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.189680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.189728 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.292194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.292234 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.292245 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.292260 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.292270 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.395337 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.395396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.395411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.395432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.395446 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.498915 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.499000 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.499025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.499058 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.499083 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.601660 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.601754 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.601765 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.601782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.601797 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.705474 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.705548 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.705569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.705597 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.705615 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.808670 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.808770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.808781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.808799 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.808810 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.815018 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.815061 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.815169 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:25 crc kubenswrapper[4837]: E1001 07:06:25.815345 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:25 crc kubenswrapper[4837]: E1001 07:06:25.815494 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:25 crc kubenswrapper[4837]: E1001 07:06:25.815652 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.912048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.912092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.912103 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.912122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:25 crc kubenswrapper[4837]: I1001 07:06:25.912136 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:25Z","lastTransitionTime":"2025-10-01T07:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.015411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.015939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.016130 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.016319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.016721 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.120499 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.120560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.120579 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.120605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.120625 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.222835 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.222891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.222902 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.222923 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.222940 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.326731 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.326803 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.326823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.326850 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.326870 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.430319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.430716 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.430884 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.431017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.431123 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.534458 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.534519 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.534537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.534563 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.534580 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.637440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.637899 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.638106 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.638311 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.638503 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.741491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.741561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.741583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.741610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.741629 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.815135 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:26 crc kubenswrapper[4837]: E1001 07:06:26.815608 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.844540 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.844605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.844623 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.844651 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.844668 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.902360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:26 crc kubenswrapper[4837]: E1001 07:06:26.902632 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:26 crc kubenswrapper[4837]: E1001 07:06:26.902731 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:34.902681943 +0000 UTC m=+51.744289408 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.948217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.948571 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.948796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.948974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:26 crc kubenswrapper[4837]: I1001 07:06:26.949121 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:26Z","lastTransitionTime":"2025-10-01T07:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.052321 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.052397 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.052422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.052449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.052466 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.155525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.155603 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.155626 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.155655 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.155677 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.258363 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.258431 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.258448 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.258471 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.258488 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.362589 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.362660 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.362682 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.362750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.362779 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.466405 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.466461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.466478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.466502 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.466519 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.569790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.570631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.570903 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.571067 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.571189 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.674562 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.675067 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.675279 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.675497 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.675669 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.779405 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.779447 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.779477 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.779494 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.779505 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.815863 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.815875 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:27 crc kubenswrapper[4837]: E1001 07:06:27.816100 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.816332 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:27 crc kubenswrapper[4837]: E1001 07:06:27.816445 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:27 crc kubenswrapper[4837]: E1001 07:06:27.816555 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.882556 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.882604 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.882621 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.882641 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.882653 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.990396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.990477 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.990501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.990531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:27 crc kubenswrapper[4837]: I1001 07:06:27.990554 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:27Z","lastTransitionTime":"2025-10-01T07:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.093133 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.093181 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.093194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.093214 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.093229 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.195942 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.195978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.195986 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.196002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.196012 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.298731 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.298798 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.298815 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.298839 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.298858 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.402189 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.402259 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.402273 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.402305 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.402323 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.505645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.505752 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.505775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.505804 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.505825 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.609518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.610453 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.610647 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.610870 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.611043 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.715375 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.715483 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.715504 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.715528 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.715545 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.815344 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:28 crc kubenswrapper[4837]: E1001 07:06:28.815526 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.818480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.818515 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.818526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.818546 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.818559 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.921840 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.921905 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.921919 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.921945 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:28 crc kubenswrapper[4837]: I1001 07:06:28.921967 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:28Z","lastTransitionTime":"2025-10-01T07:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.025052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.025112 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.025128 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.025154 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.025170 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.128082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.128127 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.128142 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.128164 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.128176 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.230243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.230325 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.230351 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.230414 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.230440 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.333543 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.333595 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.333614 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.333638 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.333655 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.436906 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.436992 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.437009 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.437038 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.437056 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.540529 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.540873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.540961 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.541095 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.541190 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.643875 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.643983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.644008 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.644035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.644056 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.747407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.747483 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.747507 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.747537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.747559 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.815951 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.816070 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.815969 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:29 crc kubenswrapper[4837]: E1001 07:06:29.816206 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:29 crc kubenswrapper[4837]: E1001 07:06:29.816383 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:29 crc kubenswrapper[4837]: E1001 07:06:29.816504 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.851217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.851256 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.851265 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.851283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.851295 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.954311 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.954368 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.954386 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.954417 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:29 crc kubenswrapper[4837]: I1001 07:06:29.954435 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:29Z","lastTransitionTime":"2025-10-01T07:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.057862 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.057939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.057961 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.057991 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.058014 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.115943 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.116023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.116048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.116082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.116107 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.137332 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:30Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.143868 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.143912 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.143922 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.143949 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.143961 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.164237 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:30Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.170599 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.170654 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.170672 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.170726 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.170747 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.194364 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:30Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.200929 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.200983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.201003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.201033 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.201059 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.227736 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:30Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.234595 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.234718 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.234751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.234789 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.234814 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.255406 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:30Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.255603 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.258636 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.258716 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.258739 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.258765 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.258784 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.363113 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.363478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.363742 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.364026 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.364258 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.469133 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.469194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.469204 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.469225 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.469237 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.572879 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.572950 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.572973 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.572999 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.573064 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.676581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.676662 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.676685 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.676750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.676774 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.780246 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.780321 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.780339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.780364 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.780382 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.815517 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:30 crc kubenswrapper[4837]: E1001 07:06:30.815745 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.817380 4837 scope.go:117] "RemoveContainer" containerID="6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.886064 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.886475 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.886489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.886508 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.886524 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.989891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.989944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.989955 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.989974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:30 crc kubenswrapper[4837]: I1001 07:06:30.989986 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:30Z","lastTransitionTime":"2025-10-01T07:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.092605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.092679 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.092721 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.092751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.092773 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.195770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.195814 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.195823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.195841 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.195853 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.218224 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/1.log" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.221867 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.222607 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.247547 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.267488 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.293051 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.305143 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.305216 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.305236 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.305264 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.305281 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.313902 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.333519 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.356066 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.375681 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.394698 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.408781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.408829 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.408838 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.408857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.408868 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.425831 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.442262 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.455030 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.470341 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.492201 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.510035 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.511969 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.512032 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.512046 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.512070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.512087 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.524083 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.540554 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:31Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.615312 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.615385 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.615411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.615439 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.615460 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.719352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.719420 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.719433 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.719459 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.719472 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.816060 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.816162 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.816226 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:31 crc kubenswrapper[4837]: E1001 07:06:31.816273 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:31 crc kubenswrapper[4837]: E1001 07:06:31.816361 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:31 crc kubenswrapper[4837]: E1001 07:06:31.816616 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.822284 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.822537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.822556 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.822589 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.822615 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.926444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.926515 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.926534 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.926565 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:31 crc kubenswrapper[4837]: I1001 07:06:31.926585 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:31Z","lastTransitionTime":"2025-10-01T07:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.029920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.029984 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.030003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.030031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.030055 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.134077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.134140 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.134159 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.134183 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.134199 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.229808 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/2.log" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.231139 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/1.log" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.236337 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.236576 4837 scope.go:117] "RemoveContainer" containerID="6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.236489 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" exitCode=1 Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.237628 4837 scope.go:117] "RemoveContainer" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" Oct 01 07:06:32 crc kubenswrapper[4837]: E1001 07:06:32.238049 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.239191 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.239258 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.239277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.239305 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.239323 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.264206 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.283297 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.298493 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.319378 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343011 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343179 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343198 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.343626 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.361044 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.374621 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.389271 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.407566 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.426364 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.441055 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.446666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.446737 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.446749 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.446771 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.446785 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.457127 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.472728 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.488521 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.505254 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.530332 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d449168ee089968e3f240370c65489667faedb05de6545b1ba65108c1855001\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:17Z\\\",\\\"message\\\":\\\" default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:17Z is after 2025-08-24T17:21:41Z]\\\\nI1001 07:06:17.378337 6242 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"58a148b3-0a7b-4412-b447-f87788c4883f\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:32Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.549820 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.549871 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.549881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.549900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.549913 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.652478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.652533 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.652545 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.652571 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.652587 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.755115 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.755163 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.755173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.755193 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.755207 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.815948 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:32 crc kubenswrapper[4837]: E1001 07:06:32.816186 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.857911 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.857985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.858004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.858034 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.858055 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.961532 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.961609 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.961627 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.961658 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:32 crc kubenswrapper[4837]: I1001 07:06:32.961678 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:32Z","lastTransitionTime":"2025-10-01T07:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.065756 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.065849 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.065876 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.065914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.065940 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.169820 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.169874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.169891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.169914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.169930 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.244996 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/2.log" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.253290 4837 scope.go:117] "RemoveContainer" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" Oct 01 07:06:33 crc kubenswrapper[4837]: E1001 07:06:33.253615 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.273348 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.273411 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.273429 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.273459 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.273478 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.275803 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.292747 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.312052 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.328419 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.350379 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.376562 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.376628 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.376644 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.376666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.376684 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.377391 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.397329 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.433428 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.459595 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.480833 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.480900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.480920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.480957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.480987 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.481317 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.499707 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.523359 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.545595 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.562606 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.580497 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.583985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.584047 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.584065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.584087 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.584101 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.600925 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.687824 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.687913 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.687928 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.687951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.687965 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.790281 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.790330 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.790341 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.790360 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.790371 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.815791 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.815859 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.815825 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:33 crc kubenswrapper[4837]: E1001 07:06:33.816000 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:33 crc kubenswrapper[4837]: E1001 07:06:33.816142 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:33 crc kubenswrapper[4837]: E1001 07:06:33.816274 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.835310 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.849138 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.868733 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.886052 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.894384 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.894449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.894464 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.894487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.894504 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.901681 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.915900 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.935074 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.952261 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.967474 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.986403 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:33Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.997860 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.997917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.997936 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.997963 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:33 crc kubenswrapper[4837]: I1001 07:06:33.997985 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:33Z","lastTransitionTime":"2025-10-01T07:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.003527 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.020440 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.035948 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.048821 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.072559 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.092101 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:34Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.100961 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.101035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.101054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.101078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.101131 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.205019 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.205091 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.205109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.205134 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.205149 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.307955 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.308013 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.308025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.308049 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.308062 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.410680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.410758 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.410770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.410792 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.410806 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.514074 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.514118 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.514130 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.514152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.514166 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.617124 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.617176 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.617186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.617202 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.617212 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.719939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.720020 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.720045 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.720071 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.720090 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.815443 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:34 crc kubenswrapper[4837]: E1001 07:06:34.815674 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.823570 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.823653 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.823678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.823754 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.823780 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.902865 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:34 crc kubenswrapper[4837]: E1001 07:06:34.903126 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:34 crc kubenswrapper[4837]: E1001 07:06:34.903235 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:06:50.903211758 +0000 UTC m=+67.744819223 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.927018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.927070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.927089 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.927114 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:34 crc kubenswrapper[4837]: I1001 07:06:34.927133 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:34Z","lastTransitionTime":"2025-10-01T07:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.004046 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.004256 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.004341 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.004388 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.004407 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:07:07.004383576 +0000 UTC m=+83.845991031 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.004497 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:07:07.004463549 +0000 UTC m=+83.846071074 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.004501 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.004618 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:07:07.004599102 +0000 UTC m=+83.846206647 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.029128 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.029170 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.029181 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.029196 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.029208 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.105394 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.105451 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105600 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105619 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105633 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105657 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105724 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105742 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105721 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:07:07.105678298 +0000 UTC m=+83.947285763 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.105831 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:07:07.105810902 +0000 UTC m=+83.947418367 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.132309 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.132355 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.132367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.132387 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.132402 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.235320 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.235408 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.235430 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.235459 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.235480 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.338587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.338666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.338684 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.338761 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.338779 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.441819 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.441877 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.441889 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.441906 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.441919 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.546102 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.546156 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.546168 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.546191 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.546204 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.650747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.650854 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.650878 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.650906 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.650927 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.754140 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.754211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.754250 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.754282 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.754303 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.815483 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.815667 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.815493 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.815818 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.815816 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:35 crc kubenswrapper[4837]: E1001 07:06:35.816009 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.857043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.857087 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.857100 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.857125 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.857149 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.960458 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.960506 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.960519 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.960537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:35 crc kubenswrapper[4837]: I1001 07:06:35.960549 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:35Z","lastTransitionTime":"2025-10-01T07:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.066011 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.066070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.066093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.066124 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.066144 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.169366 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.169435 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.169458 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.169488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.169514 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.273122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.273172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.273188 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.273209 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.273224 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.375937 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.376010 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.376034 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.376067 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.376089 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.479567 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.479652 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.479678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.479745 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.479771 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.583216 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.583274 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.583293 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.583318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.583338 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.675793 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.686907 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.687291 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.687501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.687793 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.688004 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.688153 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.705637 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.724837 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.744748 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.757645 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.785791 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.791150 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.791217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.791240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.791265 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.791285 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.805093 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.815530 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:36 crc kubenswrapper[4837]: E1001 07:06:36.815736 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.825358 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.847314 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.870805 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.892874 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.894746 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.894802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.894826 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.894857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.894880 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:36Z","lastTransitionTime":"2025-10-01T07:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.910073 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.930215 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.945489 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.960161 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.976227 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:36 crc kubenswrapper[4837]: I1001 07:06:36.992904 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:36Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.002545 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.002604 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.002630 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.002664 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.002683 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.107005 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.107071 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.107091 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.107118 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.107139 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.210345 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.210408 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.210427 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.210450 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.210469 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.313775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.313817 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.313829 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.313846 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.313858 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.418592 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.418645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.418659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.418683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.418729 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.522207 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.522240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.522249 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.522264 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.522274 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.624432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.624486 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.624494 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.624508 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.624518 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.727480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.727539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.727552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.727570 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.727584 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.815897 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:37 crc kubenswrapper[4837]: E1001 07:06:37.817030 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.817073 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:37 crc kubenswrapper[4837]: E1001 07:06:37.817239 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.817525 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:37 crc kubenswrapper[4837]: E1001 07:06:37.817745 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.830384 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.830662 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.830890 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.831063 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.831274 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.934743 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.934813 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.934834 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.934864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:37 crc kubenswrapper[4837]: I1001 07:06:37.934886 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:37Z","lastTransitionTime":"2025-10-01T07:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.038419 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.038512 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.038537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.038570 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.038595 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.142489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.142538 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.142552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.142570 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.142581 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.246643 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.246736 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.246756 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.246782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.246801 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.351604 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.351750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.351775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.351803 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.351821 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.455108 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.455151 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.455162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.455175 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.455187 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.557992 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.558050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.558061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.558077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.558086 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.661256 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.661322 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.661346 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.661374 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.661397 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.764037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.764099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.764112 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.764130 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.764145 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.815308 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:38 crc kubenswrapper[4837]: E1001 07:06:38.815526 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.868080 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.868160 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.868185 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.868217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.868248 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.971727 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.971800 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.971823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.971849 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:38 crc kubenswrapper[4837]: I1001 07:06:38.971903 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:38Z","lastTransitionTime":"2025-10-01T07:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.075659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.075769 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.075790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.075817 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.075840 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.179583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.179654 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.179673 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.179724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.179741 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.288926 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.288983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.289001 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.289024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.289039 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.392413 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.392470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.392487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.392511 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.392529 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.496098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.496167 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.496184 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.496210 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.496228 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.598822 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.598881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.598899 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.598925 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.598944 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.702372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.702444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.702463 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.702491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.702512 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.806322 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.806387 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.806408 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.806439 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.806458 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.815815 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.815876 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:39 crc kubenswrapper[4837]: E1001 07:06:39.816062 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:39 crc kubenswrapper[4837]: E1001 07:06:39.816390 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.816197 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:39 crc kubenswrapper[4837]: E1001 07:06:39.816620 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.909518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.909579 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.909599 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.909631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:39 crc kubenswrapper[4837]: I1001 07:06:39.909654 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:39Z","lastTransitionTime":"2025-10-01T07:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.013566 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.013645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.013669 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.013739 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.013761 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.117757 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.117815 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.117842 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.117890 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.117906 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.221893 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.221970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.221994 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.222023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.222043 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.275634 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.275720 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.275733 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.275751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.275764 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.298488 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:40Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.304012 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.304076 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.304099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.304125 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.304142 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.324765 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:40Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.329832 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.329873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.329890 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.329914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.329932 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.350808 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:40Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.357039 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.357100 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.357124 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.357153 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.357172 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.378341 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:40Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.383419 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.383465 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.383480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.383500 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.383512 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.403782 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:40Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.403893 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.405630 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.405717 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.405737 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.405763 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.405780 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.508991 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.509064 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.509088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.509121 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.509145 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.611795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.611861 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.611878 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.611907 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.611928 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.714944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.715045 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.715066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.715091 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.715110 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.815620 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:40 crc kubenswrapper[4837]: E1001 07:06:40.815854 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.817864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.817971 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.817986 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.818004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.818016 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.921666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.921736 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.921749 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.921771 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:40 crc kubenswrapper[4837]: I1001 07:06:40.921783 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:40Z","lastTransitionTime":"2025-10-01T07:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.024014 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.024070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.024086 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.024109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.024126 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.127929 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.127996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.128013 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.128040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.128056 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.232586 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.232676 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.232738 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.232766 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.232786 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.336855 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.336916 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.336933 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.336957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.336973 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.439566 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.439915 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.439996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.440093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.440176 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.543795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.544238 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.544485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.544765 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.545003 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.648811 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.649016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.649038 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.649065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.649085 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.752722 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.752778 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.752795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.752820 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.752841 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.815257 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.815384 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.815257 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:41 crc kubenswrapper[4837]: E1001 07:06:41.815473 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:41 crc kubenswrapper[4837]: E1001 07:06:41.815794 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:41 crc kubenswrapper[4837]: E1001 07:06:41.815965 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.855857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.855918 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.855935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.855961 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.855981 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.959434 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.959513 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.959536 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.959566 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:41 crc kubenswrapper[4837]: I1001 07:06:41.959587 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:41Z","lastTransitionTime":"2025-10-01T07:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.062627 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.062685 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.062780 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.062816 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.062846 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.167310 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.167367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.167394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.167417 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.167432 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.270980 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.271050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.271068 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.271092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.271110 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.374959 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.375029 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.375050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.375076 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.375094 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.479344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.479405 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.479422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.479446 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.479463 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.582673 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.582775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.582795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.582822 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.582844 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.687230 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.687283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.687300 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.687327 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.687345 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.791407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.791494 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.791513 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.791540 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.791558 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.815290 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:42 crc kubenswrapper[4837]: E1001 07:06:42.815484 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.894554 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.894632 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.894655 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.894683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.894731 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.997785 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.997886 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.997904 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.997932 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:42 crc kubenswrapper[4837]: I1001 07:06:42.997950 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:42Z","lastTransitionTime":"2025-10-01T07:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.101848 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.101912 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.101952 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.101987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.102011 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.205675 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.205788 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.205805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.205830 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.205847 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.308630 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.308758 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.308786 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.308815 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.308836 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.412676 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.413229 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.413396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.413547 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.413892 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.517187 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.517249 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.517267 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.517291 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.517314 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.620590 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.620645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.620658 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.620676 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.620716 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.723161 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.723201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.723211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.723228 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.723240 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.814963 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.815012 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.815008 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:43 crc kubenswrapper[4837]: E1001 07:06:43.815353 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:43 crc kubenswrapper[4837]: E1001 07:06:43.815452 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:43 crc kubenswrapper[4837]: E1001 07:06:43.815575 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.827944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.828002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.828020 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.828042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.828061 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.842419 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.860230 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.892642 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.923112 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.930287 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.930359 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.930383 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.930414 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.930441 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:43Z","lastTransitionTime":"2025-10-01T07:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.943171 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.958936 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.970628 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:43 crc kubenswrapper[4837]: I1001 07:06:43.988187 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:43Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.003459 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.016843 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033188 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033203 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033225 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033242 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.033179 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.051819 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.072402 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.088325 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.104494 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.120341 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137207 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137276 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137297 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137325 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137343 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.137523 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:44Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.239111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.239232 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.239335 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.239367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.239381 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.342217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.342254 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.342270 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.342286 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.342296 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.445201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.445260 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.445279 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.445303 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.445325 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.547746 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.547831 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.547864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.547880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.547891 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.651704 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.651753 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.651764 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.651779 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.651796 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.754829 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.754867 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.754881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.754898 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.754909 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.815902 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:44 crc kubenswrapper[4837]: E1001 07:06:44.816080 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.858330 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.858410 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.858440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.858470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.858491 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.961847 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.961910 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.961931 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.961957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:44 crc kubenswrapper[4837]: I1001 07:06:44.961975 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:44Z","lastTransitionTime":"2025-10-01T07:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.065598 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.065748 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.065774 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.065841 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.065865 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.168866 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.169327 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.169350 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.169378 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.169489 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.274078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.274146 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.274160 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.274181 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.274196 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.377628 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.377737 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.377755 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.377781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.377801 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.481306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.481364 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.481381 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.481409 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.481432 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.585896 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.585974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.585992 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.586018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.586036 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.689070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.689149 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.689174 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.689205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.689265 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.792582 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.792629 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.792644 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.792663 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.792677 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.815353 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.815423 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:45 crc kubenswrapper[4837]: E1001 07:06:45.815532 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.815832 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:45 crc kubenswrapper[4837]: E1001 07:06:45.816059 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:45 crc kubenswrapper[4837]: E1001 07:06:45.816133 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.816580 4837 scope.go:117] "RemoveContainer" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" Oct 01 07:06:45 crc kubenswrapper[4837]: E1001 07:06:45.816961 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.896084 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.896149 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.896167 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.896191 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.896210 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.999173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.999247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.999273 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.999306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:45 crc kubenswrapper[4837]: I1001 07:06:45.999348 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:45Z","lastTransitionTime":"2025-10-01T07:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.102853 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.102898 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.102914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.102937 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.102955 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.205948 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.205989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.205999 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.206016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.206027 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.308922 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.309419 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.309441 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.309468 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.309488 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.412740 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.412803 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.412824 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.412854 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.412872 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.516653 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.516751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.516771 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.516796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.516816 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.620054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.620111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.620128 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.620155 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.620175 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.723283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.723347 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.723365 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.723390 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.723408 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.815896 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:46 crc kubenswrapper[4837]: E1001 07:06:46.816074 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.826166 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.826208 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.826217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.826231 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.826245 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.930526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.930614 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.930625 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.930647 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:46 crc kubenswrapper[4837]: I1001 07:06:46.930659 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:46Z","lastTransitionTime":"2025-10-01T07:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.035275 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.035342 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.035360 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.035389 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.035407 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.138345 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.138489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.138524 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.138562 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.138586 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.242438 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.242490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.242506 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.242530 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.242547 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.345632 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.345743 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.345761 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.345784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.345801 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.448244 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.448313 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.448331 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.448352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.448371 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.552085 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.552243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.552256 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.552278 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.552292 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.655060 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.655111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.655121 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.655137 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.655147 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.757258 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.757299 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.757307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.757321 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.757330 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.815446 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:47 crc kubenswrapper[4837]: E1001 07:06:47.815607 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.815854 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:47 crc kubenswrapper[4837]: E1001 07:06:47.815913 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.816092 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:47 crc kubenswrapper[4837]: E1001 07:06:47.816296 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.859750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.859797 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.859805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.859821 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.859830 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.963355 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.963485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.963503 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.963524 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:47 crc kubenswrapper[4837]: I1001 07:06:47.963537 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:47Z","lastTransitionTime":"2025-10-01T07:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.066399 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.066457 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.066467 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.066489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.066501 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.169667 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.169760 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.169774 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.169801 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.169815 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.272492 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.272531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.272540 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.272554 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.272564 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.374789 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.374831 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.374842 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.374857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.374867 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.478157 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.478200 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.478211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.478227 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.478240 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.581802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.581900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.581928 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.581966 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.582004 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.685883 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.685970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.685989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.686018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.686236 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.789364 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.789418 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.789428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.789444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.789457 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.815166 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:48 crc kubenswrapper[4837]: E1001 07:06:48.815365 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.892472 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.892525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.892539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.892560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.892576 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.995317 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.995361 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.995370 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.995385 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:48 crc kubenswrapper[4837]: I1001 07:06:48.995393 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:48Z","lastTransitionTime":"2025-10-01T07:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.098501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.098550 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.098561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.098580 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.098591 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.201209 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.201262 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.201274 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.201347 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.201361 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.303487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.303527 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.303537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.303553 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.303563 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.406972 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.407024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.407040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.407058 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.407070 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.509538 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.509581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.509593 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.509611 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.509625 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.611897 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.611957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.611967 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.611989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.612002 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.715264 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.715298 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.715306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.715320 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.715329 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.815121 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.815201 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.815225 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:49 crc kubenswrapper[4837]: E1001 07:06:49.815282 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:49 crc kubenswrapper[4837]: E1001 07:06:49.815430 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:49 crc kubenswrapper[4837]: E1001 07:06:49.815498 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.817115 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.817148 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.817160 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.817176 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.817187 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.919465 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.919504 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.919514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.919527 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:49 crc kubenswrapper[4837]: I1001 07:06:49.919536 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:49Z","lastTransitionTime":"2025-10-01T07:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.022380 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.022454 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.022480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.022510 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.022532 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.125309 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.125377 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.125399 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.125427 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.125447 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.228752 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.228812 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.228843 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.228868 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.228885 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.331289 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.331319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.331329 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.331345 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.331356 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.434792 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.434851 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.434863 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.434881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.434894 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.472391 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.472449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.472461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.472485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.472498 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.487668 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:50Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.492116 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.492165 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.492178 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.492195 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.492208 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.507484 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:50Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.513148 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.513227 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.513244 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.513272 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.513289 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.533316 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:50Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.538928 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.539007 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.539029 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.539083 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.539106 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.559770 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:50Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.565075 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.565160 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.565180 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.565224 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.565244 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.583141 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:50Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.583374 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.585815 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.585869 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.585892 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.585914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.585935 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.688880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.688947 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.688957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.688978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.688989 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.801956 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.802013 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.802021 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.802036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.802048 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.815750 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.816055 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.905652 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.905752 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.905767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.905790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.905808 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:50Z","lastTransitionTime":"2025-10-01T07:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:50 crc kubenswrapper[4837]: I1001 07:06:50.924673 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.924922 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:50 crc kubenswrapper[4837]: E1001 07:06:50.925060 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:07:22.925014275 +0000 UTC m=+99.766621910 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.009762 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.009817 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.009835 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.009861 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.009879 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.112901 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.112993 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.113031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.113066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.113091 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.217917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.217955 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.217974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.217989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.218004 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.319485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.319559 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.319582 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.319612 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.319635 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.423289 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.423347 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.423359 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.423377 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.423394 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.526322 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.526371 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.526383 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.526400 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.526412 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.629027 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.629074 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.629083 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.629098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.629109 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.732258 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.732305 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.732316 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.732339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.732350 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.815937 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.815966 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:51 crc kubenswrapper[4837]: E1001 07:06:51.816094 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.816105 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:51 crc kubenswrapper[4837]: E1001 07:06:51.816196 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:51 crc kubenswrapper[4837]: E1001 07:06:51.816421 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.834784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.834843 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.834865 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.834894 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.834912 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.938470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.938537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.938550 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.938576 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:51 crc kubenswrapper[4837]: I1001 07:06:51.938593 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:51Z","lastTransitionTime":"2025-10-01T07:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.040873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.040925 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.040936 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.040953 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.040968 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.143562 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.143745 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.143782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.143828 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.143855 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.247407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.247466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.247481 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.247505 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.247523 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.350902 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.350969 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.350984 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.351013 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.351029 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.454635 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.454724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.454739 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.454764 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.454778 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.558338 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.558384 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.558395 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.558414 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.558425 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.662201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.662270 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.662283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.662304 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.662317 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.765382 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.765455 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.765475 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.765502 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.765525 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.814972 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:52 crc kubenswrapper[4837]: E1001 07:06:52.815171 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.868532 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.868615 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.868641 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.868678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.868747 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.971968 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.972017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.972031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.972049 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:52 crc kubenswrapper[4837]: I1001 07:06:52.972063 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:52Z","lastTransitionTime":"2025-10-01T07:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.075070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.075129 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.075139 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.075163 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.075175 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.177612 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.177663 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.177673 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.177709 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.177721 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.280943 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.281003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.281013 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.281033 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.281044 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.384110 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.384232 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.384245 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.384272 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.384282 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.487322 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.487370 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.487381 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.487398 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.487409 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.590548 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.590623 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.590645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.590675 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.590734 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.694633 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.694701 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.694711 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.694729 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.694747 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.797885 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.797967 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.797989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.798016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.798035 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.815308 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.815389 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.815330 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:53 crc kubenswrapper[4837]: E1001 07:06:53.815476 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:53 crc kubenswrapper[4837]: E1001 07:06:53.815610 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:53 crc kubenswrapper[4837]: E1001 07:06:53.815718 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.843145 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.870432 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.890922 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.900926 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.900960 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.900970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.900984 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.900995 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:53Z","lastTransitionTime":"2025-10-01T07:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.914896 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.938511 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.953371 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.970061 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:53 crc kubenswrapper[4837]: I1001 07:06:53.988237 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.000530 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:53Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.003501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.003537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.003548 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.003564 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.003576 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.013165 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.023647 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.033852 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.046256 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.060967 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.080156 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.093324 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.106247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.106298 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.106316 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.106342 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.106360 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.120093 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.209883 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.209942 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.209953 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.209971 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.209983 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.314653 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.314761 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.314872 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.314903 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.314915 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.327232 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/0.log" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.327277 4837 generic.go:334] "Generic (PLEG): container finished" podID="7dedbee2-d2c0-49fb-ac76-aa7562c61211" containerID="5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c" exitCode=1 Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.327311 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerDied","Data":"5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.327816 4837 scope.go:117] "RemoveContainer" containerID="5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.361640 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.383069 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.401577 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.417915 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.417979 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.417991 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.418014 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.418027 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.420533 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.443610 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.457475 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.472438 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.489876 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.503613 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.516915 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.520531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.520569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.520581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.520596 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.520606 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.529552 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.541683 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.556159 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.568974 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.581505 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.590999 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.611799 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:54Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.624072 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.624299 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.624391 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.624466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.624543 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.728267 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.728667 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.728880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.729073 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.729276 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.815164 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:54 crc kubenswrapper[4837]: E1001 07:06:54.815572 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.833057 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.833125 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.833143 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.833168 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.833188 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.937344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.937396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.937409 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.937428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:54 crc kubenswrapper[4837]: I1001 07:06:54.937441 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:54Z","lastTransitionTime":"2025-10-01T07:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.040542 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.040605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.040623 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.040649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.040672 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.144724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.144791 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.144808 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.144834 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.144853 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.248493 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.248560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.248585 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.248621 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.248645 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.332226 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/0.log" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.332316 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerStarted","Data":"f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.345836 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.351597 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.351656 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.351680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.351755 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.351784 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.367843 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.384933 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.402276 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.429393 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.449716 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.454363 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.454406 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.454421 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.454440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.454452 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.465275 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.477656 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.493934 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.509147 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.525228 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.536722 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.551565 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.556480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.556522 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.556534 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.556609 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.556625 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.566912 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.581647 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.593858 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.620232 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:06:55Z is after 2025-08-24T17:21:41Z" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.659678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.659735 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.659744 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.659759 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.659770 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.762767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.762829 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.762844 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.762870 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.762887 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.815399 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.815573 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:55 crc kubenswrapper[4837]: E1001 07:06:55.815774 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.815803 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:55 crc kubenswrapper[4837]: E1001 07:06:55.815953 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:55 crc kubenswrapper[4837]: E1001 07:06:55.816041 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.865782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.865821 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.865832 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.865851 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.865861 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.968252 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.968304 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.968318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.968338 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:55 crc kubenswrapper[4837]: I1001 07:06:55.968349 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:55Z","lastTransitionTime":"2025-10-01T07:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.071028 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.071080 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.071095 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.071116 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.071132 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.174143 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.174221 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.174243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.174271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.174299 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.276988 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.277600 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.277893 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.278123 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.278392 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.382186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.382266 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.382286 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.382317 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.382338 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.486310 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.486372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.486385 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.486404 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.486417 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.589111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.589169 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.589187 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.589209 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.589225 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.692977 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.693035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.693056 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.693081 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.693100 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.796186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.796284 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.796305 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.796330 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.796383 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.815159 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:56 crc kubenswrapper[4837]: E1001 07:06:56.815318 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.899509 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.899556 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.899576 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.899602 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:56 crc kubenswrapper[4837]: I1001 07:06:56.899618 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:56Z","lastTransitionTime":"2025-10-01T07:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.003524 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.003640 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.003661 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.003686 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.003742 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.106422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.106465 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.106476 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.106492 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.106504 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.209490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.209540 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.209555 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.209581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.209595 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.312473 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.312536 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.312557 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.312583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.312600 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.416048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.416105 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.416115 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.416136 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.416151 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.518987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.519033 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.519043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.519059 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.519072 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.621476 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.621541 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.621558 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.621583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.621600 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.725095 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.725172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.725196 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.725226 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.725255 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.815615 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.815683 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.815626 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:57 crc kubenswrapper[4837]: E1001 07:06:57.815911 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:57 crc kubenswrapper[4837]: E1001 07:06:57.816060 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:57 crc kubenswrapper[4837]: E1001 07:06:57.816425 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.828239 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.828277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.828290 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.828308 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.828322 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.931255 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.931309 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.931329 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.931353 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:57 crc kubenswrapper[4837]: I1001 07:06:57.931372 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:57Z","lastTransitionTime":"2025-10-01T07:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.034052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.034103 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.034119 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.034138 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.034152 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.137780 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.137849 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.137870 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.137895 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.137913 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.241427 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.241491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.241500 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.241526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.241540 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.344707 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.344767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.344784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.344811 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.344825 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.448683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.449061 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.449135 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.449221 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.449483 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.552927 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.552999 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.553024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.553054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.553076 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.656477 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.656520 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.656531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.656548 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.656561 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.759905 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.760405 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.760487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.760578 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.760654 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.815685 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:06:58 crc kubenswrapper[4837]: E1001 07:06:58.816220 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.864335 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.864824 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.864938 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.865029 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.865127 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.968615 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.969098 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.969316 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.969470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:58 crc kubenswrapper[4837]: I1001 07:06:58.969614 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:58Z","lastTransitionTime":"2025-10-01T07:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.073939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.073989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.074003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.074021 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.074036 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.177659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.177773 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.177796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.177826 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.177849 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.282093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.282444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.282539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.282632 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.282755 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.385610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.386114 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.386275 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.386416 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.386530 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.490566 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.490631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.490643 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.490702 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.490717 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.599853 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.600820 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.601066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.601255 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.601432 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.705186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.705268 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.705281 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.705301 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.705316 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.808320 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.808382 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.808394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.808419 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.808433 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.815096 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.815193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:06:59 crc kubenswrapper[4837]: E1001 07:06:59.815348 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.815124 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:06:59 crc kubenswrapper[4837]: E1001 07:06:59.815563 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:06:59 crc kubenswrapper[4837]: E1001 07:06:59.815784 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.816807 4837 scope.go:117] "RemoveContainer" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.912034 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.912074 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.912101 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.912115 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:06:59 crc kubenswrapper[4837]: I1001 07:06:59.912126 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:06:59Z","lastTransitionTime":"2025-10-01T07:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.014278 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.014633 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.014726 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.014802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.014881 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.119164 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.119219 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.119234 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.119252 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.119266 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.222066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.222112 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.222126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.222147 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.222161 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.324828 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.324868 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.324880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.324897 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.324907 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.352512 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/2.log" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.355139 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.356020 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.368673 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.383890 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.397114 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.410844 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.426018 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.427466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.427501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.427544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.427563 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.427575 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.442335 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.455941 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.469622 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.482331 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.498320 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.511016 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.523481 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.530105 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.530143 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.530155 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.530172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.530184 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.535895 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.549423 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.563982 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.579600 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.599214 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.632624 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.632681 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.632724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.632747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.632763 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.710282 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.710338 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.710352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.710374 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.710392 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.728422 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.732879 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.732920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.732934 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.732955 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.732970 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.750206 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.754796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.754848 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.754859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.754882 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.754896 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.768490 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.772307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.772383 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.772399 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.772422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.772436 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.787103 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.793873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.793927 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.793939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.793965 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.794018 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.815198 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.815354 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.817532 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:00Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:00 crc kubenswrapper[4837]: E1001 07:07:00.817854 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.819970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.820020 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.820036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.820062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.820079 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.923395 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.923453 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.923468 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.923484 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:00 crc kubenswrapper[4837]: I1001 07:07:00.923494 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:00Z","lastTransitionTime":"2025-10-01T07:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.027561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.027646 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.027667 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.027715 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.027733 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.130996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.131067 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.131079 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.131099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.131114 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.234942 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.235006 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.235024 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.235049 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.235068 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.337714 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.337767 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.337779 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.337797 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.337809 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.362188 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/3.log" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.363262 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/2.log" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.367286 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" exitCode=1 Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.367354 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.367423 4837 scope.go:117] "RemoveContainer" containerID="f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.368678 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:01 crc kubenswrapper[4837]: E1001 07:07:01.369025 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.388049 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.402223 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.416226 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.435566 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.441074 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.441271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.441395 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.441504 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.441588 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.450568 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.466468 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.485500 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.503904 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.537348 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40c17173ac152909c436698c93344cfb4df2d36030518e7c94d206c5131e630\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:31Z\\\",\\\"message\\\":\\\"er-manager-crc openshift-multus/network-metrics-daemon-z66wh openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb openshift-machine-config-operator/machine-config-daemon-lpk6v openshift-multus/multus-additional-cni-plugins-28t2p openshift-network-diagnostics/network-check-target-xd92c]\\\\nI1001 07:06:31.844460 6463 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1001 07:06:31.844487 6463 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844504 6463 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1001 07:06:31.844517 6463 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nF1001 07:06:31.844531 6463 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"07:07:00.722474 6822 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 07:07:00.722498 6822 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 07:07:00.722511 6822 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 07:07:00.722549 6822 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 07:07:00.722579 6822 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 07:07:00.722587 6822 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 07:07:00.722608 6822 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 07:07:00.722621 6822 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 07:07:00.722621 6822 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 07:07:00.722634 6822 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 07:07:00.722636 6822 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:07:00.722659 6822 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 07:07:00.722687 6822 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:07:00.722724 6822 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:07:00.722765 6822 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:07:00.723090 6822 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.544526 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.544596 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.544608 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.544631 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.544996 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.560174 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.575320 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.592832 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.619272 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.637296 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.648944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.649009 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.649026 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.649051 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.649070 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.655139 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.673438 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.695186 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:01Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.752101 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.752172 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.752189 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.752211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.752227 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.815280 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:01 crc kubenswrapper[4837]: E1001 07:07:01.815491 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.815849 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.815873 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:01 crc kubenswrapper[4837]: E1001 07:07:01.815974 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:01 crc kubenswrapper[4837]: E1001 07:07:01.816068 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.854432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.854477 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.854491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.854509 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.854523 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.957640 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.957715 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.957732 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.957751 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:01 crc kubenswrapper[4837]: I1001 07:07:01.957764 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:01Z","lastTransitionTime":"2025-10-01T07:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.060614 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.060666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.060678 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.060719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.060734 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.169444 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.169529 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.169555 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.169587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.169610 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.272453 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.272544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.272561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.272587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.272606 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.373582 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/3.log" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.375554 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.375853 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.375893 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.375914 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.375926 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.379282 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:02 crc kubenswrapper[4837]: E1001 07:07:02.379450 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.405019 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.425974 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.443515 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.463648 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.480531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.481134 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.481224 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.481283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.481303 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.481326 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.499379 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.519002 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.539810 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.551020 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.570756 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"07:07:00.722474 6822 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 07:07:00.722498 6822 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 07:07:00.722511 6822 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 07:07:00.722549 6822 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 07:07:00.722579 6822 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 07:07:00.722587 6822 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 07:07:00.722608 6822 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 07:07:00.722621 6822 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 07:07:00.722621 6822 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 07:07:00.722634 6822 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 07:07:00.722636 6822 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:07:00.722659 6822 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 07:07:00.722687 6822 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:07:00.722724 6822 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:07:00.722765 6822 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:07:00.723090 6822 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.583700 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.583732 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.583744 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.583784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.583794 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.584755 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.597343 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.612729 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.635960 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.660780 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.675139 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.686166 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.686202 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.686215 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.686253 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.686266 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.694361 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:02Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.788956 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.789003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.789015 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.789036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.789047 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.815508 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:02 crc kubenswrapper[4837]: E1001 07:07:02.815641 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.891675 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.891784 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.891814 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.891849 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.891877 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.994753 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.994821 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.994836 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.994855 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:02 crc kubenswrapper[4837]: I1001 07:07:02.994868 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:02Z","lastTransitionTime":"2025-10-01T07:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.098269 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.098329 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.098341 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.098361 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.098373 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.201992 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.202079 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.202106 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.202139 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.202160 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.306262 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.306344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.306359 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.306379 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.306397 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.409008 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.409081 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.409101 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.409126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.409144 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.511759 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.511824 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.511841 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.511866 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.511885 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.615222 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.615292 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.615307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.615337 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.615352 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.718042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.718082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.718094 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.718111 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.718122 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.815559 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.815748 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:03 crc kubenswrapper[4837]: E1001 07:07:03.815799 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.815582 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:03 crc kubenswrapper[4837]: E1001 07:07:03.815953 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:03 crc kubenswrapper[4837]: E1001 07:07:03.816064 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.822186 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.822234 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.822252 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.822303 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.822321 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.831867 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.850629 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.864613 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.880500 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.900418 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.920441 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.924522 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.924589 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.924608 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.924639 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.924663 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:03Z","lastTransitionTime":"2025-10-01T07:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.943372 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.964265 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:03 crc kubenswrapper[4837]: I1001 07:07:03.990973 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:03Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.004783 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.027612 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.027668 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.027683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.027731 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.027746 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.037468 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"07:07:00.722474 6822 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 07:07:00.722498 6822 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 07:07:00.722511 6822 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 07:07:00.722549 6822 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 07:07:00.722579 6822 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 07:07:00.722587 6822 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 07:07:00.722608 6822 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 07:07:00.722621 6822 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 07:07:00.722621 6822 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 07:07:00.722634 6822 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 07:07:00.722636 6822 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:07:00.722659 6822 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 07:07:00.722687 6822 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:07:00.722724 6822 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:07:00.722765 6822 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:07:00.723090 6822 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.061109 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.085065 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.101841 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.119103 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.130443 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.130490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.130501 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.130518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.130530 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.139139 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.156861 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:04Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.234124 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.234213 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.234240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.234277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.234300 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.337104 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.337137 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.337145 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.337160 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.337171 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.440894 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.440956 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.440974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.440999 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.441018 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.544021 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.544077 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.544090 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.544110 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.544124 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.648161 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.648236 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.648262 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.648292 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.648316 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.752298 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.752372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.752389 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.752419 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.752439 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.815425 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:04 crc kubenswrapper[4837]: E1001 07:07:04.815644 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.832060 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.855468 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.855535 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.855552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.855579 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.855599 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.958432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.958524 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.958542 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.959069 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:04 crc kubenswrapper[4837]: I1001 07:07:04.959296 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:04Z","lastTransitionTime":"2025-10-01T07:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.063062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.063132 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.063142 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.063164 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.063179 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.166241 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.166320 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.166339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.166365 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.166385 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.269243 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.269306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.269325 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.269350 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.269376 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.372790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.372857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.372870 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.372890 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.372902 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.476557 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.476614 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.476624 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.476645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.476657 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.581840 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.581910 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.581928 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.581968 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.581986 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.685776 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.685820 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.685830 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.685847 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.685857 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.789237 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.789297 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.789318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.789343 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.789361 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.815261 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.815272 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:05 crc kubenswrapper[4837]: E1001 07:07:05.815479 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.815498 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:05 crc kubenswrapper[4837]: E1001 07:07:05.815624 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:05 crc kubenswrapper[4837]: E1001 07:07:05.815804 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.894580 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.894647 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.894666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.894724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.894741 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.998214 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.998251 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.998262 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.998280 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:05 crc kubenswrapper[4837]: I1001 07:07:05.998292 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:05Z","lastTransitionTime":"2025-10-01T07:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.101921 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.101969 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.101980 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.101998 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.102010 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.205487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.205537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.205553 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.205578 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.205597 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.308805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.308892 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.308913 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.308939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.308958 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.413018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.413082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.413094 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.413114 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.413128 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.516970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.517037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.517048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.517066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.517075 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.626265 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.626975 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.627016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.627052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.627079 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.730360 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.730423 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.730436 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.730457 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.730471 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.815638 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:06 crc kubenswrapper[4837]: E1001 07:07:06.815970 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.834679 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.834769 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.834795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.834825 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.834845 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.938199 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.938258 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.938294 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.938324 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:06 crc kubenswrapper[4837]: I1001 07:07:06.938344 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:06Z","lastTransitionTime":"2025-10-01T07:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.018533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.018848 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.018927 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.018992 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.018956785 +0000 UTC m=+147.860564280 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.019079 4837 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.019140 4837 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.019178 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.01915171 +0000 UTC m=+147.860759205 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.019251 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.019220702 +0000 UTC m=+147.860828197 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.041567 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.041609 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.041621 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.041639 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.041651 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.120015 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.120092 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120320 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120349 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120370 4837 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120375 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120424 4837 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120450 4837 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120453 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.120430971 +0000 UTC m=+147.962038456 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.120560 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.120531363 +0000 UTC m=+147.962138858 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.145149 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.145211 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.145235 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.145263 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.145285 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.248977 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.249040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.249064 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.249092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.249113 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.352686 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.352783 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.352800 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.352826 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.352843 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.456277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.456376 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.456401 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.456438 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.456469 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.559307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.559394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.559408 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.559426 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.559439 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.662481 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.662527 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.662557 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.662578 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.662590 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.765825 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.765900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.765924 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.765951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.765968 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.815804 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.815939 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.815804 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.816027 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.816226 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:07 crc kubenswrapper[4837]: E1001 07:07:07.816398 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.869256 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.869316 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.869649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.869862 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.870070 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.972895 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.972958 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.972976 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.973003 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:07 crc kubenswrapper[4837]: I1001 07:07:07.973021 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:07Z","lastTransitionTime":"2025-10-01T07:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.076621 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.076683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.076723 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.076747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.076765 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.180194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.180270 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.180290 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.180318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.180338 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.284555 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.285416 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.285436 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.285470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.285549 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.389518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.389592 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.389609 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.389634 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.389647 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.492782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.492865 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.492889 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.492918 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.492938 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.602895 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.602981 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.602994 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.603016 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.603031 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.707290 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.707379 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.707396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.707422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.707441 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.810791 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.810843 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.810859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.810888 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.810906 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.815491 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:08 crc kubenswrapper[4837]: E1001 07:07:08.815684 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.914806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.914860 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.914877 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.914901 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:08 crc kubenswrapper[4837]: I1001 07:07:08.914919 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:08Z","lastTransitionTime":"2025-10-01T07:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.018177 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.018235 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.018252 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.018275 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.018293 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.121277 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.121339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.121360 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.121389 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.121411 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.224404 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.224470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.224488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.224516 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.224533 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.328614 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.328666 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.328682 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.328747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.328766 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.432392 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.432449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.432466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.432490 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.432510 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.537179 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.537261 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.537282 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.537307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.537325 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.650920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.650978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.650995 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.651018 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.651039 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.754437 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.754731 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.754929 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.755088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.755240 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.815462 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:09 crc kubenswrapper[4837]: E1001 07:07:09.815611 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.815462 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.815727 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:09 crc kubenswrapper[4837]: E1001 07:07:09.815939 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:09 crc kubenswrapper[4837]: E1001 07:07:09.816026 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.858196 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.858523 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.858681 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.858861 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.859016 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.961947 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.962037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.962047 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.962068 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:09 crc kubenswrapper[4837]: I1001 07:07:09.962082 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:09Z","lastTransitionTime":"2025-10-01T07:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.065214 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.065301 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.065329 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.065364 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.065389 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.168919 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.168971 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.168983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.169002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.169015 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.271950 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.272023 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.272044 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.272068 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.272086 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.374822 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.374880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.374890 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.374908 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.374918 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.478205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.478246 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.478254 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.478271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.478280 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.581822 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.581910 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.581949 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.581983 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.582010 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.685478 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.685557 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.685580 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.685612 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.685633 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.793806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.793874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.793892 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.793919 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.793936 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.815318 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.815507 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.860461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.860531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.860548 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.860573 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.860590 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.882322 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.888917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.888987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.889004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.889030 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.889050 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.911296 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.917014 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.917069 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.917092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.917122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.917143 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.937734 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.943620 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.943717 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.943742 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.943773 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.943794 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.963402 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.969719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.969787 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.969805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.969832 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.969851 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.990908 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:10Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:10 crc kubenswrapper[4837]: E1001 07:07:10.991135 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.994021 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.994075 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.994085 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.994108 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:10 crc kubenswrapper[4837]: I1001 07:07:10.994122 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:10Z","lastTransitionTime":"2025-10-01T07:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.097782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.097838 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.097850 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.097873 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.097886 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.201162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.201222 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.201240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.201264 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.201280 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.304425 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.304473 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.304483 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.304504 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.304516 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.407911 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.407966 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.407982 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.408006 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.408026 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.511546 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.511606 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.511628 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.511658 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.511681 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.615008 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.615109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.615131 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.615157 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.615177 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.718543 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.718620 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.718642 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.718680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.718740 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.815167 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:11 crc kubenswrapper[4837]: E1001 07:07:11.815367 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.815361 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.815408 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:11 crc kubenswrapper[4837]: E1001 07:07:11.815861 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:11 crc kubenswrapper[4837]: E1001 07:07:11.815919 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.823507 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.823558 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.823569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.823588 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.823601 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.926426 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.926499 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.926525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.926551 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:11 crc kubenswrapper[4837]: I1001 07:07:11.926567 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:11Z","lastTransitionTime":"2025-10-01T07:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.036450 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.036502 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.036515 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.036535 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.036547 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.140336 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.140396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.140407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.140424 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.140435 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.244387 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.244448 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.244461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.244484 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.244502 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.348640 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.348768 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.348796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.348827 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.348846 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.451623 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.451756 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.451786 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.451817 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.451840 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.554443 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.554505 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.554521 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.554545 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.554574 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.658120 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.658200 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.658223 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.658254 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.658279 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.762555 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.762618 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.762634 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.762654 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.762668 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.815618 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:12 crc kubenswrapper[4837]: E1001 07:07:12.815881 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.865489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.865583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.865636 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.865661 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.865679 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.969183 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.969247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.969268 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.969293 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:12 crc kubenswrapper[4837]: I1001 07:07:12.969310 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:12Z","lastTransitionTime":"2025-10-01T07:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.072658 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.072737 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.072757 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.072776 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.072789 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.175899 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.175963 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.175990 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.176025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.176050 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.280293 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.280383 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.280409 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.280440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.280463 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.384201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.384292 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.384318 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.384394 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.384419 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.488594 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.488659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.488679 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.488752 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.488778 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.592455 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.592579 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.592605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.592643 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.592668 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.696065 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.696126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.696137 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.696156 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.696168 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.799839 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.799888 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.799901 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.799922 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.799940 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.815780 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:13 crc kubenswrapper[4837]: E1001 07:07:13.815956 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.816041 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.816072 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:13 crc kubenswrapper[4837]: E1001 07:07:13.816408 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:13 crc kubenswrapper[4837]: E1001 07:07:13.816788 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.834268 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.860387 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"07:07:00.722474 6822 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 07:07:00.722498 6822 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 07:07:00.722511 6822 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 07:07:00.722549 6822 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 07:07:00.722579 6822 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 07:07:00.722587 6822 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 07:07:00.722608 6822 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 07:07:00.722621 6822 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 07:07:00.722621 6822 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 07:07:00.722634 6822 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 07:07:00.722636 6822 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:07:00.722659 6822 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 07:07:00.722687 6822 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:07:00.722724 6822 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:07:00.722765 6822 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:07:00.723090 6822 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.884982 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.902985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.903036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.903056 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.903078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.903091 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:13Z","lastTransitionTime":"2025-10-01T07:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.905684 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.929112 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.958682 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.978881 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:13 crc kubenswrapper[4837]: I1001 07:07:13.996834 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:13Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.007052 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.007109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.007126 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.007155 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.007175 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.017660 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.039209 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.059431 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.076494 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.093894 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.110349 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.111274 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.111302 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.111311 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.111326 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.111346 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.124417 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.141937 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.155497 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7e6d663-c20c-4863-9212-7965d24ad655\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58bf5812f5a56f60cf83e3908a227ea3bf593692379d9371c27b51d03b93f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.178083 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:14Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.214447 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.214497 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.214514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.214538 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.214557 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.318205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.318306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.318602 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.318926 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.318978 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.422515 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.422570 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.422587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.422611 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.422633 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.526422 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.526482 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.526498 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.526521 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.526538 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.629401 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.629454 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.629466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.629485 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.629497 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.732688 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.732781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.732798 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.732827 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.732847 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.816892 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:14 crc kubenswrapper[4837]: E1001 07:07:14.817537 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.817611 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:14 crc kubenswrapper[4837]: E1001 07:07:14.817986 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.836112 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.836169 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.836194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.836222 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.836242 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.939664 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.939746 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.939760 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.939783 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:14 crc kubenswrapper[4837]: I1001 07:07:14.939797 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:14Z","lastTransitionTime":"2025-10-01T07:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.043305 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.043372 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.043396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.043428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.043451 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.145845 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.145934 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.145954 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.145985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.146005 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.249322 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.249397 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.249432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.249464 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.249486 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.353683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.353842 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.353870 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.353904 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.353931 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.457352 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.457449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.457476 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.457512 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.457536 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.560165 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.560207 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.560215 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.560228 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.560241 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.662421 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.662484 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.662507 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.662549 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.662571 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.766542 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.766625 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.766645 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.766671 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.766726 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.815573 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:15 crc kubenswrapper[4837]: E1001 07:07:15.815815 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.815944 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:15 crc kubenswrapper[4837]: E1001 07:07:15.816540 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.816854 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:15 crc kubenswrapper[4837]: E1001 07:07:15.817139 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.870391 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.870469 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.870491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.870520 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.870538 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.974307 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.974382 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.974402 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.974429 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:15 crc kubenswrapper[4837]: I1001 07:07:15.974447 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:15Z","lastTransitionTime":"2025-10-01T07:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.077344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.077407 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.077424 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.077446 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.077461 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.180440 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.180508 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.180525 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.180547 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.180565 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.283493 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.283552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.283561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.283576 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.283588 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.386809 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.386886 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.386902 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.386925 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.386941 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.489782 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.489834 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.489847 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.489864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.489876 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.593783 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.593902 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.593932 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.593960 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.593979 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.697194 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.697257 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.697270 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.697296 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.697313 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.801171 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.801260 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.801287 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.801319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.801345 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.815367 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:16 crc kubenswrapper[4837]: E1001 07:07:16.815535 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.904559 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.904610 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.904622 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.904641 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:16 crc kubenswrapper[4837]: I1001 07:07:16.904652 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:16Z","lastTransitionTime":"2025-10-01T07:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.007487 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.007563 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.007580 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.007607 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.007627 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.111326 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.111409 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.111433 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.111464 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.111487 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.214824 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.214950 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.214978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.215073 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.215101 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.318174 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.318247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.318266 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.318291 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.318311 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.421908 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.422007 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.422019 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.422043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.422056 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.524981 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.525040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.525050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.525066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.525077 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.629632 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.629711 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.629728 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.629747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.629760 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.732858 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.732912 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.732932 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.732955 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.732972 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.815046 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.815229 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:17 crc kubenswrapper[4837]: E1001 07:07:17.815307 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:17 crc kubenswrapper[4837]: E1001 07:07:17.815506 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.815061 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:17 crc kubenswrapper[4837]: E1001 07:07:17.815686 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.835847 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.835906 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.835935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.835966 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.835989 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.938843 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.938896 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.938912 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.938930 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:17 crc kubenswrapper[4837]: I1001 07:07:17.938942 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:17Z","lastTransitionTime":"2025-10-01T07:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.041957 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.041988 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.041996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.042012 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.042022 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.145225 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.145303 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.145323 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.145350 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.145437 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.248427 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.248496 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.248514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.248539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.248554 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.357175 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.357335 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.358093 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.358125 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.358146 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.460988 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.461048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.461066 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.461092 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.461113 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.564974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.565017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.565035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.565063 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.565087 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.668747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.669195 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.669366 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.669514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.669742 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.772826 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.772869 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.772882 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.772898 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.772910 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.816349 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:18 crc kubenswrapper[4837]: E1001 07:07:18.816736 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.877175 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.877229 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.877240 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.877262 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.877277 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.979913 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.979966 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.979985 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.980007 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:18 crc kubenswrapper[4837]: I1001 07:07:18.980024 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:18Z","lastTransitionTime":"2025-10-01T07:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.082795 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.083514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.083587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.083662 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.084009 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.187216 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.187287 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.187310 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.187342 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.187368 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.290681 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.290774 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.290792 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.290818 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.290839 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.393747 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.393834 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.393852 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.393880 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.393902 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.496598 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.496679 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.496750 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.496785 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.496852 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.599665 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.599771 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.599796 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.599864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.599886 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.702577 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.702633 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.702649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.702671 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.702708 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.806259 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.806306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.806319 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.806339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.806354 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.815794 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.815850 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.816154 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:19 crc kubenswrapper[4837]: E1001 07:07:19.816321 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:19 crc kubenswrapper[4837]: E1001 07:07:19.816495 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:19 crc kubenswrapper[4837]: E1001 07:07:19.816735 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.909396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.909470 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.909488 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.909516 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:19 crc kubenswrapper[4837]: I1001 07:07:19.909534 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:19Z","lastTransitionTime":"2025-10-01T07:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.013469 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.013537 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.013556 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.013585 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.013604 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.116717 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.116766 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.116777 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.116797 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.116810 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.220349 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.220412 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.220434 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.220461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.220479 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.324450 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.324514 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.324533 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.324560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.324584 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.440339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.440396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.440409 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.440432 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.440454 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.543218 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.543606 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.543710 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.543805 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.543915 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.647920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.647996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.648015 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.648042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.648064 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.751806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.751859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.751874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.751896 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.751914 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.815282 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:20 crc kubenswrapper[4837]: E1001 07:07:20.815503 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.854900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.855070 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.855099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.855131 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.855154 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.959071 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.959153 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.959178 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.959209 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:20 crc kubenswrapper[4837]: I1001 07:07:20.959232 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:20Z","lastTransitionTime":"2025-10-01T07:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.063640 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.063745 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.063764 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.063793 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.063811 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.167270 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.167323 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.167340 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.167367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.167383 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.272924 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.273017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.273037 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.273063 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.273081 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.310509 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.310569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.310596 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.310625 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.310650 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.329882 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:21Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.336231 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.336310 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.336339 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.336366 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.336389 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.357581 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:21Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.362638 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.362845 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.362944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.363033 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.363137 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.381761 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:21Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.386560 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.386618 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.386635 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.386659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.386677 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.403577 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:21Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.407958 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.408034 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.408059 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.408084 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.408102 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.424488 4837 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T07:07:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dd6f714d-42ba-4793-b8d3-f7eb0219041e\\\",\\\"systemUUID\\\":\\\"1fa2a216-c629-41b0-a072-88b5c711fdce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:21Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.424867 4837 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.426775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.426823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.426840 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.426864 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.426881 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.530919 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.531323 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.531397 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.531491 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.531575 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.634552 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.634896 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.634974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.635050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.635124 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.738205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.738258 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.738278 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.738306 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.738325 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.815432 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.815499 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.815443 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.815624 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.815777 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:21 crc kubenswrapper[4837]: E1001 07:07:21.815858 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.841215 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.841600 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.841685 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.841806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.841880 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.945616 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.945719 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.945739 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.945763 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:21 crc kubenswrapper[4837]: I1001 07:07:21.945781 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:21Z","lastTransitionTime":"2025-10-01T07:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.048585 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.048649 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.048675 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.048745 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.048772 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.152521 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.152571 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.152590 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.152615 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.152634 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.260903 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.261583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.262054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.262272 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.262407 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.365989 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.366035 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.366048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.366062 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.366071 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.469414 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.469493 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.469518 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.469547 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.469568 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.572915 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.572967 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.572987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.573011 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.573028 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.675677 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.675785 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.675806 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.675830 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.675847 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.780486 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.780534 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.780544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.780561 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.780574 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.815016 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:22 crc kubenswrapper[4837]: E1001 07:07:22.815391 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.883731 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.883803 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.883823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.883850 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.883873 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.987105 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.987177 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.987197 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.987225 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:22 crc kubenswrapper[4837]: I1001 07:07:22.987244 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:22Z","lastTransitionTime":"2025-10-01T07:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.015188 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:23 crc kubenswrapper[4837]: E1001 07:07:23.015414 4837 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:07:23 crc kubenswrapper[4837]: E1001 07:07:23.015559 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs podName:95803be9-8a27-4569-958d-81666ad8defc nodeName:}" failed. No retries permitted until 2025-10-01 07:08:27.015527042 +0000 UTC m=+163.857134497 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs") pod "network-metrics-daemon-z66wh" (UID: "95803be9-8a27-4569-958d-81666ad8defc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.091392 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.091780 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.091857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.091924 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.091993 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.195592 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.195654 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.195672 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.195723 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.195742 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.299497 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.299538 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.299547 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.299569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.299580 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.403971 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.404040 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.404060 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.404088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.404108 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.508252 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.508312 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.508329 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.508349 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.508366 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.611770 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.611920 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.611946 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.611973 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.611991 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.715954 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.716006 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.716019 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.716043 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.716059 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.815945 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.815945 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.816017 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:23 crc kubenswrapper[4837]: E1001 07:07:23.816291 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:23 crc kubenswrapper[4837]: E1001 07:07:23.816437 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:23 crc kubenswrapper[4837]: E1001 07:07:23.816559 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.818857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.818970 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.819050 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.819539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.819624 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.834139 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de5eda3a31480c3f5acda0763332b55fc55939cef8c2837dbf50c356367a513\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.849432 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48ebd466-35e3-41ad-9c3d-c7703c2286c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1186f63efdaf10a2063d5198b4d496bb15613e203ef8202dba1f9f51dbd005ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06a040b8f76bf19cc0ea4dd3b5868253037ce2353428062b2bea0cbb8a45073e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0e5148c3f81060a6932c875d39170d83f6636b8627e3db09fcedfc145a6f765\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc9c35f6db3057bd94bc077588042097414fddc7093ca0a5d420439a7535b2c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.866560 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7e6d663-c20c-4863-9212-7965d24ad655\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58bf5812f5a56f60cf83e3908a227ea3bf593692379d9371c27b51d03b93f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a871065b81cf9fb545c7fa396b35e12cd9b30f997e08fb7265f8fe3a222244e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.886449 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a4b24c-e0ac-4976-9105-d605fb6f5843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84cab92aeca565c94bb9ab5484798fde360ba88100207522b45abf5ac4ad5885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ed61897dad83092b144d126e88dd19dfaa7e9202c3300f75a2e6bd4d685bfd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7901358b190d3f6f94cb8b856cdf237831a54ffbdbd1e967a8d4746560adb69e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://beca73b2f37915e280395d0c383608b2ff69e3a75889e636f10f53e7c5f35539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09dd208de61731d08ba3d18d1d4e217cc2bcede9e0c29c293ba6e1b4e31a6028\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T07:05:57Z\\\",\\\"message\\\":\\\"W1001 07:05:47.091209 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1001 07:05:47.091604 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759302347 cert, and key in /tmp/serving-cert-3482622033/serving-signer.crt, /tmp/serving-cert-3482622033/serving-signer.key\\\\nI1001 07:05:47.378290 1 observer_polling.go:159] Starting file observer\\\\nW1001 07:05:47.383221 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1001 07:05:47.383461 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 07:05:47.386786 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3482622033/tls.crt::/tmp/serving-cert-3482622033/tls.key\\\\\\\"\\\\nF1001 07:05:57.643167 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c51ae7e94d862991743a8cecdd9e3132e2a0a8e0b173e0248543c4a4d63382a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ec8a8386f6b7892cea4bb853dfafe755523ce2c1d68dd9caf2a53be8810100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.900640 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnhdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d89cf1-3c27-4aa7-ae3f-5e9284fbf555\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7897b13832bc15bf6dc5bb8ac0360feeff7b44fea6ed61de64e1b1cf6ab590d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmx25\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnhdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.922153 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"401d9e9e-cba5-413c-b078-83858883db16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:07:00Z\\\",\\\"message\\\":\\\"07:07:00.722474 6822 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 07:07:00.722498 6822 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 07:07:00.722511 6822 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 07:07:00.722549 6822 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 07:07:00.722579 6822 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1001 07:07:00.722587 6822 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1001 07:07:00.722608 6822 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 07:07:00.722621 6822 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1001 07:07:00.722621 6822 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 07:07:00.722634 6822 handler.go:208] Removed *v1.Node event handler 7\\\\nI1001 07:07:00.722636 6822 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 07:07:00.722659 6822 handler.go:208] Removed *v1.Node event handler 2\\\\nI1001 07:07:00.722687 6822 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 07:07:00.722724 6822 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 07:07:00.722765 6822 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 07:07:00.723090 6822 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5btcx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vf2cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.928326 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.928379 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.928396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.928448 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.928469 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:23Z","lastTransitionTime":"2025-10-01T07:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.950929 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-28t2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9533dad3-ca00-4500-bfcb-614a5ee75545\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4eed03a52c76c43b4a4412cec525291ff4ee1ad391a119cd4fb552b3cf735c46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5bc37193f1ccbd5de87fc1562bf087d3620f555250f54819fd8a7eccc1ed7ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a96be16291533d6f642df17cbdbfab11b637675d8d892844c39bd48c37242cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a286b37e636545cc433b7f5deae7d63260c9e52eb23ebc94a52b52dc6f3b9a61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad43f8caff5af89dbc0a93b3df2800bacb37c197ea364d2510b0557659a7aa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89ecc81d1d38e4924eae6424e0442d2966bd42bf50aeb0e75d6989109c65aadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db8941ed89f59675e2c05ddde604ec6e0f33acac0a10f84474b0e17c9eafadec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zv8f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-28t2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.964492 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-kztv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dedbee2-d2c0-49fb-ac76-aa7562c61211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T07:06:53Z\\\",\\\"message\\\":\\\"2025-10-01T07:06:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29\\\\n2025-10-01T07:06:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_afab5765-8d8c-4d98-92aa-7bb8c2fdca29 to /host/opt/cni/bin/\\\\n2025-10-01T07:06:08Z [verbose] multus-daemon started\\\\n2025-10-01T07:06:08Z [verbose] Readiness Indicator file check\\\\n2025-10-01T07:06:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ddpts\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-kztv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.976075 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zdszv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a30c54-f521-4faf-9a65-0e85b3ebe8c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04d8f671f6b12c1049c98b49d86929237fe010d052c882ef64a189ab585dd3ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sn4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zdszv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:23 crc kubenswrapper[4837]: I1001 07:07:23.995738 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:23Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.012458 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2097b0e4c28ed06aa2def23c0be570965bd97eccbd5285dec8129c5d42564ae9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50607a7f86d1bb7ce610c497fde350a7bf9463bcb31078f7ca961c0a0e29193f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.030077 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a57550826949dc1e62946d871c05f0bb1947dd68284f4079d3b367760f43d038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.033058 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.033131 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.033162 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.033197 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.033222 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.049203 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e12bb16-7807-437f-927c-71dc8d0a27ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f30613d11a5150e551c1ee2711ff23ec9f0d30d4774171efb915721bf60bc111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5m2hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lpk6v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.069504 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09e0351f-93e5-44bc-803c-5a722722f722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cf5992f2096c2ee3b249d9dd3b9243ef43e597e2bb60cab60e76183ae6b4991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdcea5dad68038ab2452775db3dfdf5e03ba174accdbff325bd6e2beb50979c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f47sj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s9qpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.085465 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z66wh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95803be9-8a27-4569-958d-81666ad8defc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfs6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:06:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z66wh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.100806 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"976f67df-e552-4231-b1df-0a51fa4bcba6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T07:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74567a09f8ba8edab0a0be1a94a144b35c5b2fb0614ae208a85d9d3ece1575bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ff263185d91ec8e1d1b5687440266b5130729763c7ee9046431efc878c584b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbee45f87f5cb6d34261b638e7f3ea4f4839dac0b69c937c6c4d7a527a42c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T07:05:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fb35202070cbcc0e6cd0dbc3c8994d3bbcccb332937d6615271292c8f9ff4f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T07:05:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T07:05:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T07:05:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.119542 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.136356 4837 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T07:06:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T07:07:24Z is after 2025-08-24T17:21:41Z" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.137288 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.137338 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.137357 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.137382 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.137401 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.240031 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.240094 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.240117 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.240161 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.240187 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.343461 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.343520 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.343545 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.343575 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.343634 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.446683 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.446801 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.446828 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.446857 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.446875 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.550801 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.550874 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.550891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.550915 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.550930 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.654541 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.654594 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.654605 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.654626 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.654643 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.757995 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.758064 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.758083 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.758109 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.758129 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.815753 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:24 crc kubenswrapper[4837]: E1001 07:07:24.816014 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.861720 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.861793 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.861813 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.861840 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.861861 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.965953 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.966027 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.966048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.966079 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:24 crc kubenswrapper[4837]: I1001 07:07:24.966102 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:24Z","lastTransitionTime":"2025-10-01T07:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.070193 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.070271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.070300 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.070334 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.070359 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.173519 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.173587 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.173607 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.173630 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.173649 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.277068 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.277152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.277175 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.277205 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.277226 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.381179 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.381248 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.381267 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.381293 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.381311 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.483929 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.483987 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.484001 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.484025 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.484047 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.587353 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.587428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.587448 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.587476 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.587496 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.690363 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.690417 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.690429 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.690449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.690463 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.793975 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.794036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.794054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.794076 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.794092 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.817823 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:25 crc kubenswrapper[4837]: E1001 07:07:25.817980 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.818190 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:25 crc kubenswrapper[4837]: E1001 07:07:25.818241 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.818351 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:25 crc kubenswrapper[4837]: E1001 07:07:25.818396 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.897361 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.897429 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.897449 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.897476 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:25 crc kubenswrapper[4837]: I1001 07:07:25.897495 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:25Z","lastTransitionTime":"2025-10-01T07:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.000837 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.000891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.000910 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.000935 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.000953 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.105058 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.105146 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.105171 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.105201 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.105219 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.207887 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.207956 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.207974 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.208001 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.208023 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.310991 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.311048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.311058 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.311075 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.311088 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.414289 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.414355 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.414371 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.414396 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.414411 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.517799 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.517876 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.517893 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.517917 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.517932 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.621638 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.621724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.621736 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.621757 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.621769 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.724668 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.724749 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.724764 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.724785 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.724797 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.815828 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:26 crc kubenswrapper[4837]: E1001 07:07:26.816092 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.828002 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.828091 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.828119 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.828159 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.828193 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.836029 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.932489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.932568 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.932581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.932602 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:26 crc kubenswrapper[4837]: I1001 07:07:26.932615 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:26Z","lastTransitionTime":"2025-10-01T07:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.036015 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.036078 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.036088 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.036108 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.036122 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.138540 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.138583 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.138592 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.138608 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.138620 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.240819 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.240900 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.240918 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.240944 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.240962 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.343311 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.343351 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.343360 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.343377 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.343387 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.445530 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.445588 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.445599 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.445617 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.445629 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.548581 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.548634 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.548644 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.548663 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.548673 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.652099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.652169 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.652193 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.652224 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.652247 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.755344 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.755388 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.755404 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.755428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.755445 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.815960 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.816101 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.816156 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:27 crc kubenswrapper[4837]: E1001 07:07:27.816389 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:27 crc kubenswrapper[4837]: E1001 07:07:27.816486 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:27 crc kubenswrapper[4837]: E1001 07:07:27.816676 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.817491 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:27 crc kubenswrapper[4837]: E1001 07:07:27.817652 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.859351 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.859523 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.859604 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.859636 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.859744 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.963290 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.963367 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.963388 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.963466 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:27 crc kubenswrapper[4837]: I1001 07:07:27.963491 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:27Z","lastTransitionTime":"2025-10-01T07:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.066418 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.066486 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.066504 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.066531 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.066554 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.169584 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.169659 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.169677 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.169741 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.169762 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.273371 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.273447 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.273459 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.273479 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.273489 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.376369 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.376421 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.376433 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.376454 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.376466 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.480223 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.480317 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.480331 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.480357 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.480369 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.583135 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.583187 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.583197 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.583218 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.583231 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.687149 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.687217 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.687239 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.687265 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.687284 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.790441 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.790539 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.790564 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.790604 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.790635 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.814987 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:28 crc kubenswrapper[4837]: E1001 07:07:28.815228 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.893846 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.893906 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.893923 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.893950 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.893967 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.997489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.997569 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.997589 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.997620 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:28 crc kubenswrapper[4837]: I1001 07:07:28.997639 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:28Z","lastTransitionTime":"2025-10-01T07:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.100368 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.100428 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.100439 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.100456 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.100469 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.203486 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.203527 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.203535 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.203549 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.203558 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.306939 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.306996 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.307019 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.307048 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.307070 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.410342 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.410421 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.410450 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.410489 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.410517 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.514090 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.514152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.514173 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.514196 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.514216 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.617284 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.617365 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.617390 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.617426 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.617450 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.721042 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.721113 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.721132 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.721157 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.721176 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.815592 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.815737 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.815840 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:29 crc kubenswrapper[4837]: E1001 07:07:29.815902 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:29 crc kubenswrapper[4837]: E1001 07:07:29.816071 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:29 crc kubenswrapper[4837]: E1001 07:07:29.816254 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.824769 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.824839 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.824859 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.824886 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.824914 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.928054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.928135 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.928154 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.928183 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:29 crc kubenswrapper[4837]: I1001 07:07:29.928202 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:29Z","lastTransitionTime":"2025-10-01T07:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.031790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.031869 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.031888 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.031913 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.031930 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.135891 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.135964 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.135981 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.136008 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.136026 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.239819 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.240054 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.240131 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.240158 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.240183 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.344017 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.344082 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.344099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.344122 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.344141 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.447624 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.447722 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.447733 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.447772 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.447789 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.551724 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.551780 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.551790 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.551813 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.551824 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.655754 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.655823 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.655848 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.655881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.655899 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.759086 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.759154 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.759177 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.759202 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.759220 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.815609 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:30 crc kubenswrapper[4837]: E1001 07:07:30.816231 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.862038 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.862099 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.862116 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.862141 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.862162 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.966176 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.966586 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.966792 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.966951 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:30 crc kubenswrapper[4837]: I1001 07:07:30.967079 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:30Z","lastTransitionTime":"2025-10-01T07:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.070881 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.070930 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.070948 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.070973 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.070993 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.174448 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.174507 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.174521 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.174544 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.174557 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.277913 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.277978 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.278004 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.278036 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.278058 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.381174 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.381234 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.381251 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.381283 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.381308 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.485152 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.485236 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.485247 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.485271 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.485285 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.589404 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.589480 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.589680 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.589746 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.589767 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.693706 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.693781 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.693799 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.693828 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.693878 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.754595 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.754746 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.754775 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.754802 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.754820 4837 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T07:07:31Z","lastTransitionTime":"2025-10-01T07:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.816042 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.816164 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.816337 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:31 crc kubenswrapper[4837]: E1001 07:07:31.816904 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:31 crc kubenswrapper[4837]: E1001 07:07:31.816613 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:31 crc kubenswrapper[4837]: E1001 07:07:31.817133 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.834302 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md"] Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.834908 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.838677 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.839055 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.839668 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.840175 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.921147 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-kztv6" podStartSLOduration=87.921122844 podStartE2EDuration="1m27.921122844s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:31.920726995 +0000 UTC m=+108.762334470" watchObservedRunningTime="2025-10-01 07:07:31.921122844 +0000 UTC m=+108.762730309" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.921358 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-28t2p" podStartSLOduration=87.92135231 podStartE2EDuration="1m27.92135231s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:31.902597384 +0000 UTC m=+108.744204849" watchObservedRunningTime="2025-10-01 07:07:31.92135231 +0000 UTC m=+108.762959775" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.927779 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52475579-8c65-4adf-8817-04c4fc58ffae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.927838 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.927899 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52475579-8c65-4adf-8817-04c4fc58ffae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.927933 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52475579-8c65-4adf-8817-04c4fc58ffae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.927981 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:31 crc kubenswrapper[4837]: I1001 07:07:31.933271 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zdszv" podStartSLOduration=87.933259443 podStartE2EDuration="1m27.933259443s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:31.932862443 +0000 UTC m=+108.774469908" watchObservedRunningTime="2025-10-01 07:07:31.933259443 +0000 UTC m=+108.774866908" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.027250 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podStartSLOduration=88.027229438 podStartE2EDuration="1m28.027229438s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.010320129 +0000 UTC m=+108.851927594" watchObservedRunningTime="2025-10-01 07:07:32.027229438 +0000 UTC m=+108.868836903" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.027591 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s9qpb" podStartSLOduration=87.027586587 podStartE2EDuration="1m27.027586587s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.026619663 +0000 UTC m=+108.868227128" watchObservedRunningTime="2025-10-01 07:07:32.027586587 +0000 UTC m=+108.869194052" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029513 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52475579-8c65-4adf-8817-04c4fc58ffae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029579 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029624 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52475579-8c65-4adf-8817-04c4fc58ffae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029648 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029721 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52475579-8c65-4adf-8817-04c4fc58ffae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.029735 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.030628 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/52475579-8c65-4adf-8817-04c4fc58ffae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.031009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52475579-8c65-4adf-8817-04c4fc58ffae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.043603 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52475579-8c65-4adf-8817-04c4fc58ffae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.059249 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52475579-8c65-4adf-8817-04c4fc58ffae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ml7md\" (UID: \"52475579-8c65-4adf-8817-04c4fc58ffae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.087384 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.087363115 podStartE2EDuration="56.087363115s" podCreationTimestamp="2025-10-01 07:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.073317359 +0000 UTC m=+108.914924824" watchObservedRunningTime="2025-10-01 07:07:32.087363115 +0000 UTC m=+108.928970580" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.130363 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.130332176 podStartE2EDuration="28.130332176s" podCreationTimestamp="2025-10-01 07:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.107054765 +0000 UTC m=+108.948662230" watchObservedRunningTime="2025-10-01 07:07:32.130332176 +0000 UTC m=+108.971939671" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.163729 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.202078 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.202047137 podStartE2EDuration="1m25.202047137s" podCreationTimestamp="2025-10-01 07:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.200537409 +0000 UTC m=+109.042144864" watchObservedRunningTime="2025-10-01 07:07:32.202047137 +0000 UTC m=+109.043654632" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.202263 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.202255122 podStartE2EDuration="6.202255122s" podCreationTimestamp="2025-10-01 07:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.172091966 +0000 UTC m=+109.013699471" watchObservedRunningTime="2025-10-01 07:07:32.202255122 +0000 UTC m=+109.043862617" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.265408 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.265383415 podStartE2EDuration="1m31.265383415s" podCreationTimestamp="2025-10-01 07:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.265141279 +0000 UTC m=+109.106748754" watchObservedRunningTime="2025-10-01 07:07:32.265383415 +0000 UTC m=+109.106990890" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.281954 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tnhdf" podStartSLOduration=88.281929915 podStartE2EDuration="1m28.281929915s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.280475078 +0000 UTC m=+109.122082583" watchObservedRunningTime="2025-10-01 07:07:32.281929915 +0000 UTC m=+109.123537380" Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.501136 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" event={"ID":"52475579-8c65-4adf-8817-04c4fc58ffae","Type":"ContainerStarted","Data":"689d9087329d8306726968fa31ad3db53803c93423cbdebeaebdd396e5cdbea1"} Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.501234 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" event={"ID":"52475579-8c65-4adf-8817-04c4fc58ffae","Type":"ContainerStarted","Data":"118c97471e76c0a5f31b7827800b66e16a3183af2be9a0c9e1292c7982b9bc03"} Oct 01 07:07:32 crc kubenswrapper[4837]: I1001 07:07:32.815113 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:32 crc kubenswrapper[4837]: E1001 07:07:32.815317 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:33 crc kubenswrapper[4837]: I1001 07:07:33.815932 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:33 crc kubenswrapper[4837]: I1001 07:07:33.816028 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:33 crc kubenswrapper[4837]: E1001 07:07:33.818860 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:33 crc kubenswrapper[4837]: I1001 07:07:33.818985 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:33 crc kubenswrapper[4837]: E1001 07:07:33.819034 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:33 crc kubenswrapper[4837]: E1001 07:07:33.819205 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:34 crc kubenswrapper[4837]: I1001 07:07:34.815219 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:34 crc kubenswrapper[4837]: E1001 07:07:34.815814 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:35 crc kubenswrapper[4837]: I1001 07:07:35.815993 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:35 crc kubenswrapper[4837]: E1001 07:07:35.816122 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:35 crc kubenswrapper[4837]: I1001 07:07:35.816214 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:35 crc kubenswrapper[4837]: E1001 07:07:35.816369 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:35 crc kubenswrapper[4837]: I1001 07:07:35.816222 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:35 crc kubenswrapper[4837]: E1001 07:07:35.816482 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:36 crc kubenswrapper[4837]: I1001 07:07:36.815114 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:36 crc kubenswrapper[4837]: E1001 07:07:36.815333 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:37 crc kubenswrapper[4837]: I1001 07:07:37.815668 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:37 crc kubenswrapper[4837]: I1001 07:07:37.815757 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:37 crc kubenswrapper[4837]: E1001 07:07:37.816755 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:37 crc kubenswrapper[4837]: I1001 07:07:37.815801 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:37 crc kubenswrapper[4837]: E1001 07:07:37.817226 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:37 crc kubenswrapper[4837]: E1001 07:07:37.817499 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:38 crc kubenswrapper[4837]: I1001 07:07:38.816205 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:38 crc kubenswrapper[4837]: E1001 07:07:38.817334 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:38 crc kubenswrapper[4837]: I1001 07:07:38.818138 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:38 crc kubenswrapper[4837]: E1001 07:07:38.818481 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vf2cm_openshift-ovn-kubernetes(401d9e9e-cba5-413c-b078-83858883db16)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" Oct 01 07:07:39 crc kubenswrapper[4837]: I1001 07:07:39.815969 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:39 crc kubenswrapper[4837]: I1001 07:07:39.816095 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:39 crc kubenswrapper[4837]: I1001 07:07:39.816235 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:39 crc kubenswrapper[4837]: E1001 07:07:39.816866 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:39 crc kubenswrapper[4837]: E1001 07:07:39.817179 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:39 crc kubenswrapper[4837]: E1001 07:07:39.817586 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.534321 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/1.log" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.535141 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/0.log" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.535377 4837 generic.go:334] "Generic (PLEG): container finished" podID="7dedbee2-d2c0-49fb-ac76-aa7562c61211" containerID="f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952" exitCode=1 Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.535488 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerDied","Data":"f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952"} Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.535779 4837 scope.go:117] "RemoveContainer" containerID="5d7109516c291331799154c853e0b549e2e6b85bc5d7be886b3bca380864014c" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.536330 4837 scope.go:117] "RemoveContainer" containerID="f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952" Oct 01 07:07:40 crc kubenswrapper[4837]: E1001 07:07:40.536600 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-kztv6_openshift-multus(7dedbee2-d2c0-49fb-ac76-aa7562c61211)\"" pod="openshift-multus/multus-kztv6" podUID="7dedbee2-d2c0-49fb-ac76-aa7562c61211" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.566484 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ml7md" podStartSLOduration=96.56645778 podStartE2EDuration="1m36.56645778s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:32.524246537 +0000 UTC m=+109.365854022" watchObservedRunningTime="2025-10-01 07:07:40.56645778 +0000 UTC m=+117.408065265" Oct 01 07:07:40 crc kubenswrapper[4837]: I1001 07:07:40.815777 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:40 crc kubenswrapper[4837]: E1001 07:07:40.816168 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:41 crc kubenswrapper[4837]: I1001 07:07:41.542128 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/1.log" Oct 01 07:07:41 crc kubenswrapper[4837]: I1001 07:07:41.815408 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:41 crc kubenswrapper[4837]: I1001 07:07:41.815614 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:41 crc kubenswrapper[4837]: E1001 07:07:41.815814 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:41 crc kubenswrapper[4837]: I1001 07:07:41.815941 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:41 crc kubenswrapper[4837]: E1001 07:07:41.816180 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:41 crc kubenswrapper[4837]: E1001 07:07:41.816263 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:42 crc kubenswrapper[4837]: I1001 07:07:42.815144 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:42 crc kubenswrapper[4837]: E1001 07:07:42.815363 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:43 crc kubenswrapper[4837]: I1001 07:07:43.815268 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:43 crc kubenswrapper[4837]: E1001 07:07:43.817514 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:43 crc kubenswrapper[4837]: I1001 07:07:43.817595 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:43 crc kubenswrapper[4837]: I1001 07:07:43.817784 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:43 crc kubenswrapper[4837]: E1001 07:07:43.817970 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:43 crc kubenswrapper[4837]: E1001 07:07:43.818169 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:43 crc kubenswrapper[4837]: E1001 07:07:43.828871 4837 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 01 07:07:43 crc kubenswrapper[4837]: E1001 07:07:43.934014 4837 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:07:44 crc kubenswrapper[4837]: I1001 07:07:44.815825 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:44 crc kubenswrapper[4837]: E1001 07:07:44.816033 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:45 crc kubenswrapper[4837]: I1001 07:07:45.815712 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:45 crc kubenswrapper[4837]: I1001 07:07:45.815896 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:45 crc kubenswrapper[4837]: E1001 07:07:45.815927 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:45 crc kubenswrapper[4837]: I1001 07:07:45.816017 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:45 crc kubenswrapper[4837]: E1001 07:07:45.816174 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:45 crc kubenswrapper[4837]: E1001 07:07:45.816211 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:46 crc kubenswrapper[4837]: I1001 07:07:46.815723 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:46 crc kubenswrapper[4837]: E1001 07:07:46.815874 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:47 crc kubenswrapper[4837]: I1001 07:07:47.815985 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:47 crc kubenswrapper[4837]: I1001 07:07:47.816021 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:47 crc kubenswrapper[4837]: I1001 07:07:47.815990 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:47 crc kubenswrapper[4837]: E1001 07:07:47.816115 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:47 crc kubenswrapper[4837]: E1001 07:07:47.816277 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:47 crc kubenswrapper[4837]: E1001 07:07:47.816264 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:48 crc kubenswrapper[4837]: I1001 07:07:48.815894 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:48 crc kubenswrapper[4837]: E1001 07:07:48.816067 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:48 crc kubenswrapper[4837]: E1001 07:07:48.935128 4837 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:07:49 crc kubenswrapper[4837]: I1001 07:07:49.818505 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:49 crc kubenswrapper[4837]: I1001 07:07:49.818535 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:49 crc kubenswrapper[4837]: E1001 07:07:49.819778 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:49 crc kubenswrapper[4837]: E1001 07:07:49.819941 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:49 crc kubenswrapper[4837]: I1001 07:07:49.818707 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:49 crc kubenswrapper[4837]: E1001 07:07:49.820294 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:50 crc kubenswrapper[4837]: I1001 07:07:50.815165 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:50 crc kubenswrapper[4837]: E1001 07:07:50.815358 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:51 crc kubenswrapper[4837]: I1001 07:07:51.815217 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:51 crc kubenswrapper[4837]: I1001 07:07:51.815216 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:51 crc kubenswrapper[4837]: I1001 07:07:51.815217 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:51 crc kubenswrapper[4837]: E1001 07:07:51.817316 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:51 crc kubenswrapper[4837]: E1001 07:07:51.817571 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:51 crc kubenswrapper[4837]: E1001 07:07:51.817587 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:52 crc kubenswrapper[4837]: I1001 07:07:52.815350 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:52 crc kubenswrapper[4837]: E1001 07:07:52.815563 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:53 crc kubenswrapper[4837]: I1001 07:07:53.815561 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:53 crc kubenswrapper[4837]: I1001 07:07:53.817420 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:53 crc kubenswrapper[4837]: E1001 07:07:53.817416 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:53 crc kubenswrapper[4837]: I1001 07:07:53.817478 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:53 crc kubenswrapper[4837]: E1001 07:07:53.817743 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:53 crc kubenswrapper[4837]: E1001 07:07:53.818091 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:53 crc kubenswrapper[4837]: I1001 07:07:53.819304 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:07:53 crc kubenswrapper[4837]: E1001 07:07:53.935870 4837 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.592158 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/3.log" Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.595363 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerStarted","Data":"5b0c10fc9ea9b7cbb60afef0c27d6ccfed8c99619eb8bc8ba7a28d700b7e31b9"} Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.596107 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.629811 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podStartSLOduration=109.629789706 podStartE2EDuration="1m49.629789706s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:07:54.625803405 +0000 UTC m=+131.467410880" watchObservedRunningTime="2025-10-01 07:07:54.629789706 +0000 UTC m=+131.471397181" Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.685564 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z66wh"] Oct 01 07:07:54 crc kubenswrapper[4837]: I1001 07:07:54.689214 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:54 crc kubenswrapper[4837]: E1001 07:07:54.689450 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:55 crc kubenswrapper[4837]: I1001 07:07:55.815818 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:55 crc kubenswrapper[4837]: I1001 07:07:55.815927 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:55 crc kubenswrapper[4837]: I1001 07:07:55.815937 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:55 crc kubenswrapper[4837]: E1001 07:07:55.816096 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:55 crc kubenswrapper[4837]: E1001 07:07:55.816654 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:55 crc kubenswrapper[4837]: E1001 07:07:55.816807 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:55 crc kubenswrapper[4837]: I1001 07:07:55.817293 4837 scope.go:117] "RemoveContainer" containerID="f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952" Oct 01 07:07:56 crc kubenswrapper[4837]: I1001 07:07:56.606835 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/1.log" Oct 01 07:07:56 crc kubenswrapper[4837]: I1001 07:07:56.607553 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerStarted","Data":"efa528d87556ee38ed9f50eab56f31e7f63b2e759fdc058103598ba4f496886e"} Oct 01 07:07:56 crc kubenswrapper[4837]: I1001 07:07:56.815146 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:56 crc kubenswrapper[4837]: E1001 07:07:56.815351 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:57 crc kubenswrapper[4837]: I1001 07:07:57.815951 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:57 crc kubenswrapper[4837]: I1001 07:07:57.816075 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:57 crc kubenswrapper[4837]: E1001 07:07:57.816190 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 07:07:57 crc kubenswrapper[4837]: I1001 07:07:57.816378 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:57 crc kubenswrapper[4837]: E1001 07:07:57.816476 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 07:07:57 crc kubenswrapper[4837]: E1001 07:07:57.816645 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 07:07:58 crc kubenswrapper[4837]: I1001 07:07:58.815177 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:07:58 crc kubenswrapper[4837]: E1001 07:07:58.815417 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z66wh" podUID="95803be9-8a27-4569-958d-81666ad8defc" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.815271 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.816957 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.818062 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.823817 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.823867 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.823845 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 07:07:59 crc kubenswrapper[4837]: I1001 07:07:59.823985 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 07:08:00 crc kubenswrapper[4837]: I1001 07:08:00.815544 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:08:00 crc kubenswrapper[4837]: I1001 07:08:00.817321 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 07:08:00 crc kubenswrapper[4837]: I1001 07:08:00.818966 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.435988 4837 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.496199 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-822f9"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.496786 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.498686 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rklvs"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.499726 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.501897 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.502228 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.503923 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.505011 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.510650 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.511297 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.513273 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.513722 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.520418 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.520721 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.520982 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.521201 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.521382 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.521843 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.522859 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.523516 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.524177 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkxdr"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.525001 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn9dw"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.525241 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.526100 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.526170 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.527994 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.528188 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.528315 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.528477 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.529156 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.529326 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.531256 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.532317 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.543581 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.544507 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.548615 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.548835 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.548990 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549211 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549438 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549535 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549598 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549726 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.549932 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-25t4p"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.550046 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.550188 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.550231 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.550445 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.551644 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.551981 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.551986 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.552133 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.552313 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.552463 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.553035 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.553150 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.553236 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.553329 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.557580 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.558577 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.558799 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.558831 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.559219 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.559248 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.559734 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.560028 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.560833 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.559750 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.560306 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.560896 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.561322 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.561442 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.564136 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.564240 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.564269 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.564356 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.564524 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565057 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565079 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565342 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565395 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565412 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565346 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565351 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565514 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565786 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565866 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565882 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565978 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.565998 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.566005 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.566080 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.566932 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.566986 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.567113 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.567312 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.567319 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.567391 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.568778 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.569189 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.569318 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.569427 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.569705 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.570044 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.570187 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.570367 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.570483 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.570874 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.571018 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.572048 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.572128 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.575881 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.576433 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xszsl"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.576912 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.577212 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.601131 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.601386 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.601594 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.601760 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.601916 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.602074 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.603090 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.603246 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.610858 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.610922 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611133 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611312 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611350 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611493 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611500 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611723 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611790 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611822 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvl6\" (UniqueName: \"kubernetes.io/projected/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-kube-api-access-xhvl6\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611843 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611894 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/968df061-3925-4301-9444-1b02e679de29-serving-cert\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611935 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611959 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.611987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612072 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612106 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612142 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-image-import-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612173 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit-dir\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612197 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/db418233-a22c-4134-8414-93c3bbb3bf1f-machine-approver-tls\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612218 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwxhp\" (UniqueName: \"kubernetes.io/projected/5be4a157-1e4f-4e93-8f85-2c01e1f60ff1-kube-api-access-cwxhp\") pod \"downloads-7954f5f757-25t4p\" (UID: \"5be4a157-1e4f-4e93-8f85-2c01e1f60ff1\") " pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612242 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hm7\" (UniqueName: \"kubernetes.io/projected/67a8d1d3-d463-4e79-be09-1e552f2fde05-kube-api-access-q5hm7\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612261 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612304 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612350 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sl9v\" (UniqueName: \"kubernetes.io/projected/d5baef47-d279-4a19-88f6-271f843d6594-kube-api-access-5sl9v\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612374 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvftf\" (UniqueName: \"kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612401 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612421 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612443 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5kjp\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-kube-api-access-d5kjp\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612462 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612477 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-node-pullsecrets\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612516 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612535 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-serving-cert\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612554 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612575 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612613 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-serving-cert\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612636 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612655 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612673 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612722 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612745 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612767 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612789 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612811 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-config\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5baef47-d279-4a19-88f6-271f843d6594-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612855 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612877 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-client\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612915 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgcds\" (UniqueName: \"kubernetes.io/projected/968df061-3925-4301-9444-1b02e679de29-kube-api-access-tgcds\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612936 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612974 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.612993 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613015 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2abd0aad-2266-4b79-9eb0-f03154ed176c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613035 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-trusted-ca\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613059 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-auth-proxy-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613076 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvvmf\" (UniqueName: \"kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613104 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znr77\" (UniqueName: \"kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613127 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-encryption-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613154 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613174 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-images\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613193 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnkkj\" (UniqueName: \"kubernetes.io/projected/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-kube-api-access-hnkkj\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613254 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2abd0aad-2266-4b79-9eb0-f03154ed176c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613276 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613298 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613323 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-config\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613369 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wft6\" (UniqueName: \"kubernetes.io/projected/db418233-a22c-4134-8414-93c3bbb3bf1f-kube-api-access-9wft6\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613391 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67a8d1d3-d463-4e79-be09-1e552f2fde05-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613412 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613434 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqk4x\" (UniqueName: \"kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613461 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-config\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.613822 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.614575 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.615763 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.619120 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.619427 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.644928 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.645066 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.645641 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.646566 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.647112 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.647557 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.648351 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.648477 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.648930 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.649506 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-smwgx"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.652465 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.652917 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.652982 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.653011 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-vb728"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.653095 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.653331 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.653806 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.657208 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.657832 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.660348 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.661927 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.662314 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.662554 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.673653 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rwwps"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.674066 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.674446 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.674814 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.674876 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.675039 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.675113 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.675451 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.676454 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rlk6w"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.677028 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.677352 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.678384 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.678876 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.679402 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.679642 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.680600 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.680956 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.681111 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.681514 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.682064 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.682643 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.683299 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.683466 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.685041 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.685190 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.686161 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.689204 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rklvs"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.691843 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.691904 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.694615 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn9dw"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.696471 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hjphq"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.699634 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.701051 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.701812 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.702136 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-822f9"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.703797 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.713308 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714714 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714754 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714796 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-image-import-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714820 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714837 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit-dir\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714853 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zpz\" (UniqueName: \"kubernetes.io/projected/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-kube-api-access-z7zpz\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714869 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-policies\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714885 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lwgq\" (UniqueName: \"kubernetes.io/projected/4f6c496c-0b1e-4e04-a70c-56ae687b3503-kube-api-access-5lwgq\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714902 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hm7\" (UniqueName: \"kubernetes.io/projected/67a8d1d3-d463-4e79-be09-1e552f2fde05-kube-api-access-q5hm7\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714919 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714938 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/db418233-a22c-4134-8414-93c3bbb3bf1f-machine-approver-tls\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.714960 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwxhp\" (UniqueName: \"kubernetes.io/projected/5be4a157-1e4f-4e93-8f85-2c01e1f60ff1-kube-api-access-cwxhp\") pod \"downloads-7954f5f757-25t4p\" (UID: \"5be4a157-1e4f-4e93-8f85-2c01e1f60ff1\") " pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715001 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715021 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715042 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-service-ca-bundle\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715068 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sl9v\" (UniqueName: \"kubernetes.io/projected/d5baef47-d279-4a19-88f6-271f843d6594-kube-api-access-5sl9v\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715088 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvftf\" (UniqueName: \"kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715108 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715167 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715195 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-trusted-ca\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715222 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715242 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n5v4\" (UniqueName: \"kubernetes.io/projected/6cdca78e-93e0-4e36-94a6-51a65c42ef75-kube-api-access-7n5v4\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715270 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5kjp\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-kube-api-access-d5kjp\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715313 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-node-pullsecrets\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715341 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-client\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715364 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-service-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715403 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715426 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-serving-cert\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715468 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715488 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715507 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715545 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715561 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-serving-cert\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715577 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715595 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715613 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715638 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715655 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9hhs\" (UniqueName: \"kubernetes.io/projected/9b4e607c-15af-4ba5-8501-eb50c38c522d-kube-api-access-p9hhs\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715672 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715709 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-config\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715727 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5baef47-d279-4a19-88f6-271f843d6594-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715744 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715787 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-client\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-config\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715825 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e771b-41ef-4984-aacb-6dd2e1231815-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715872 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgcds\" (UniqueName: \"kubernetes.io/projected/968df061-3925-4301-9444-1b02e679de29-kube-api-access-tgcds\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715948 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715971 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jpnn\" (UniqueName: \"kubernetes.io/projected/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-kube-api-access-4jpnn\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.715987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716025 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716043 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-metrics-certs\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716060 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-serving-cert\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716100 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-client\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716119 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716137 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716178 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716197 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716214 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2abd0aad-2266-4b79-9eb0-f03154ed176c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716229 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-trusted-ca\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716248 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b4e607c-15af-4ba5-8501-eb50c38c522d-metrics-tls\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716267 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716286 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwljd\" (UniqueName: \"kubernetes.io/projected/065e771b-41ef-4984-aacb-6dd2e1231815-kube-api-access-fwljd\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716304 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wkd2\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-kube-api-access-5wkd2\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716323 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvvmf\" (UniqueName: \"kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716341 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znr77\" (UniqueName: \"kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716357 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-auth-proxy-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716373 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-encryption-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716389 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-images\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnkkj\" (UniqueName: \"kubernetes.io/projected/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-kube-api-access-hnkkj\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716458 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-serving-cert\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716477 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716492 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716517 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-config\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716533 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716552 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2abd0aad-2266-4b79-9eb0-f03154ed176c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716577 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716604 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716800 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-config\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716880 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716908 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-serving-cert\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-default-certificate\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716973 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065e771b-41ef-4984-aacb-6dd2e1231815-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.716990 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-metrics-tls\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717035 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67a8d1d3-d463-4e79-be09-1e552f2fde05-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717074 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqk4x\" (UniqueName: \"kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717117 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717140 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wft6\" (UniqueName: \"kubernetes.io/projected/db418233-a22c-4134-8414-93c3bbb3bf1f-kube-api-access-9wft6\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-stats-auth\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717208 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-dir\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717227 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717270 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717293 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-config\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717313 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvl6\" (UniqueName: \"kubernetes.io/projected/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-kube-api-access-xhvl6\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717379 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717423 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/968df061-3925-4301-9444-1b02e679de29-serving-cert\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717453 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717470 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717514 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-encryption-config\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717536 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717556 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbp4r\" (UniqueName: \"kubernetes.io/projected/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-kube-api-access-dbp4r\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.717602 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-config\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.719038 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.720595 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit-dir\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.721093 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-config\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.721107 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.721256 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.721326 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-node-pullsecrets\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.721519 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rwwps"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.722671 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.722804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-image-import-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.724095 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2abd0aad-2266-4b79-9eb0-f03154ed176c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.724500 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.725081 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.725978 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-images\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.726153 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.726650 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xszsl"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.726996 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db418233-a22c-4134-8414-93c3bbb3bf1f-auth-proxy-config\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.727171 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.727213 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.727611 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-trusted-ca\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.727615 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-service-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.727833 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.728437 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.728515 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.728766 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.729041 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.729511 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.730064 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5baef47-d279-4a19-88f6-271f843d6594-config\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.730454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.730454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.730859 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.731446 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2abd0aad-2266-4b79-9eb0-f03154ed176c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.731801 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.731936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.732028 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.732115 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.732132 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/db418233-a22c-4134-8414-93c3bbb3bf1f-machine-approver-tls\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.732153 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.732117 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733088 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733366 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733435 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733474 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733628 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733729 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.733974 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.734116 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.734454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/968df061-3925-4301-9444-1b02e679de29-config\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.734752 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.735297 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-encryption-config\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.735440 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.735529 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.736072 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-audit\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.736288 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-etcd-client\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.736367 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/67a8d1d3-d463-4e79-be09-1e552f2fde05-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.736536 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-serving-cert\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.737472 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.737547 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-smwgx"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.738763 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-serving-cert\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.739045 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.739192 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5baef47-d279-4a19-88f6-271f843d6594-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.740736 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.745782 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-25t4p"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.748726 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.750111 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkxdr"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.751083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/968df061-3925-4301-9444-1b02e679de29-serving-cert\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.751177 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.752189 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.753284 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.754539 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.755563 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-m78s5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.757211 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.757338 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.758025 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.758880 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.759150 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.760229 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.761919 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.762624 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9lclf"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.764590 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rlk6w"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.764730 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.764931 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.765884 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.766876 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.767864 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m78s5"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.768959 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9lclf"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.770849 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-45m2c"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.771417 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-45m2c"] Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.771501 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.779919 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.798963 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818188 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-config\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818432 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lwgq\" (UniqueName: \"kubernetes.io/projected/4f6c496c-0b1e-4e04-a70c-56ae687b3503-kube-api-access-5lwgq\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818527 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zpz\" (UniqueName: \"kubernetes.io/projected/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-kube-api-access-z7zpz\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818776 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-policies\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.818993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-service-ca-bundle\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-trusted-ca\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n5v4\" (UniqueName: \"kubernetes.io/projected/6cdca78e-93e0-4e36-94a6-51a65c42ef75-kube-api-access-7n5v4\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-client\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819504 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-service-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819587 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819777 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.819904 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-policies\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.820208 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9hhs\" (UniqueName: \"kubernetes.io/projected/9b4e607c-15af-4ba5-8501-eb50c38c522d-kube-api-access-p9hhs\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.820371 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-config\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.820540 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e771b-41ef-4984-aacb-6dd2e1231815-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.820397 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-service-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.820261 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821022 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821286 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jpnn\" (UniqueName: \"kubernetes.io/projected/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-kube-api-access-4jpnn\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821412 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-metrics-certs\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821737 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-serving-cert\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821888 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-client\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822041 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.821216 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-config\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822276 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b4e607c-15af-4ba5-8501-eb50c38c522d-metrics-tls\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822414 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822534 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwljd\" (UniqueName: \"kubernetes.io/projected/065e771b-41ef-4984-aacb-6dd2e1231815-kube-api-access-fwljd\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wkd2\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-kube-api-access-5wkd2\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.822095 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.823371 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.823554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.823995 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-serving-cert\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824157 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824323 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-config\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824467 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824994 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-metrics-tls\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.825142 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-serving-cert\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.825280 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-default-certificate\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.825510 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065e771b-41ef-4984-aacb-6dd2e1231815-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824926 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-config\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824926 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.825287 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-client\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.824730 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f6c496c-0b1e-4e04-a70c-56ae687b3503-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.825669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.826136 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-stats-auth\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.826339 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-dir\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.826500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827191 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827016 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-serving-cert\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827126 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6cdca78e-93e0-4e36-94a6-51a65c42ef75-etcd-ca\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827218 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-serving-cert\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.826452 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f6c496c-0b1e-4e04-a70c-56ae687b3503-audit-dir\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827774 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-encryption-config\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827940 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbp4r\" (UniqueName: \"kubernetes.io/projected/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-kube-api-access-dbp4r\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.827867 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cdca78e-93e0-4e36-94a6-51a65c42ef75-serving-cert\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.828223 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-etcd-client\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.831213 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4f6c496c-0b1e-4e04-a70c-56ae687b3503-encryption-config\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.860273 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.879719 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.899257 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.920020 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.927646 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.939144 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.941085 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.960582 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.979305 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 01 07:08:02 crc kubenswrapper[4837]: I1001 07:08:02.999081 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.002051 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065e771b-41ef-4984-aacb-6dd2e1231815-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.019828 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.039386 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.050565 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065e771b-41ef-4984-aacb-6dd2e1231815-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.059899 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.080017 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.090421 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-metrics-tls\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.107668 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.111199 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-trusted-ca\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.119250 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.140484 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.158678 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.179595 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.200160 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.207340 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.220116 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.229989 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.239732 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.249819 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9b4e607c-15af-4ba5-8501-eb50c38c522d-metrics-tls\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.260458 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.279437 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.299228 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.320530 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.338931 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.351232 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.360354 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.379936 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.390436 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-config\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.399741 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.420433 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.431026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-default-certificate\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.440223 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.447167 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-metrics-certs\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.459793 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.470121 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-stats-auth\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.487866 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.499200 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.500811 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-service-ca-bundle\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.519075 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.560113 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.579431 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.599842 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.620220 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.640399 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.660603 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.677382 4837 request.go:700] Waited for 1.001437588s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.680534 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.699673 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.718610 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.770888 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.771054 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.779792 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.800282 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.820367 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.840326 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.859965 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.879739 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.899614 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.920044 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.940089 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.959637 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.989401 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 01 07:08:03 crc kubenswrapper[4837]: I1001 07:08:03.998844 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.020076 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.040178 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.062636 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.079243 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.100617 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.120081 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.139470 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.160812 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.179744 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.200184 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.220658 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.240992 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.260048 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.278905 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.300056 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.319343 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.340092 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.390549 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.407466 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hm7\" (UniqueName: \"kubernetes.io/projected/67a8d1d3-d463-4e79-be09-1e552f2fde05-kube-api-access-q5hm7\") pod \"cluster-samples-operator-665b6dd947-tx9ss\" (UID: \"67a8d1d3-d463-4e79-be09-1e552f2fde05\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.418312 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgcds\" (UniqueName: \"kubernetes.io/projected/968df061-3925-4301-9444-1b02e679de29-kube-api-access-tgcds\") pod \"authentication-operator-69f744f599-mkxdr\" (UID: \"968df061-3925-4301-9444-1b02e679de29\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.444937 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znr77\" (UniqueName: \"kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77\") pod \"console-f9d7485db-mg4h6\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.458182 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnkkj\" (UniqueName: \"kubernetes.io/projected/785b6d63-389e-4eb8-ba3b-67f23d5dd3c3-kube-api-access-hnkkj\") pod \"apiserver-76f77b778f-rklvs\" (UID: \"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3\") " pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.463123 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.481421 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvvmf\" (UniqueName: \"kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf\") pod \"oauth-openshift-558db77b4-mlw9m\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.486325 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.495708 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvftf\" (UniqueName: \"kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf\") pod \"controller-manager-879f6c89f-4cx5v\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.503759 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.515028 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5kjp\" (UniqueName: \"kubernetes.io/projected/2abd0aad-2266-4b79-9eb0-f03154ed176c-kube-api-access-d5kjp\") pod \"cluster-image-registry-operator-dc59b4c8b-9pf94\" (UID: \"2abd0aad-2266-4b79-9eb0-f03154ed176c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.542333 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wft6\" (UniqueName: \"kubernetes.io/projected/db418233-a22c-4134-8414-93c3bbb3bf1f-kube-api-access-9wft6\") pod \"machine-approver-56656f9798-mdg4v\" (UID: \"db418233-a22c-4134-8414-93c3bbb3bf1f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.557850 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwxhp\" (UniqueName: \"kubernetes.io/projected/5be4a157-1e4f-4e93-8f85-2c01e1f60ff1-kube-api-access-cwxhp\") pod \"downloads-7954f5f757-25t4p\" (UID: \"5be4a157-1e4f-4e93-8f85-2c01e1f60ff1\") " pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.586036 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sl9v\" (UniqueName: \"kubernetes.io/projected/d5baef47-d279-4a19-88f6-271f843d6594-kube-api-access-5sl9v\") pod \"machine-api-operator-5694c8668f-822f9\" (UID: \"d5baef47-d279-4a19-88f6-271f843d6594\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.600969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqk4x\" (UniqueName: \"kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x\") pod \"route-controller-manager-6576b87f9c-t8vmh\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.613885 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.623266 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.624123 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvl6\" (UniqueName: \"kubernetes.io/projected/485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7-kube-api-access-xhvl6\") pod \"console-operator-58897d9998-cn9dw\" (UID: \"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7\") " pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.628179 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.639097 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.659621 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.677453 4837 request.go:700] Waited for 1.912430171s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.677882 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.687056 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.693258 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss"] Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.698778 4837 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.709843 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.719209 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.734268 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mkxdr"] Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.740332 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 07:08:04 crc kubenswrapper[4837]: W1001 07:08:04.742655 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod968df061_3925_4301_9444_1b02e679de29.slice/crio-9e4dcf6d83c6d11c07805b28e63a1272cc5ee0e27b032371fd5ac5146f8d1af5 WatchSource:0}: Error finding container 9e4dcf6d83c6d11c07805b28e63a1272cc5ee0e27b032371fd5ac5146f8d1af5: Status 404 returned error can't find the container with id 9e4dcf6d83c6d11c07805b28e63a1272cc5ee0e27b032371fd5ac5146f8d1af5 Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.748264 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.760152 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.770951 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.780503 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.781957 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.792293 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.799670 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.813954 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.828128 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.845109 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lwgq\" (UniqueName: \"kubernetes.io/projected/4f6c496c-0b1e-4e04-a70c-56ae687b3503-kube-api-access-5lwgq\") pod \"apiserver-7bbb656c7d-bftpg\" (UID: \"4f6c496c-0b1e-4e04-a70c-56ae687b3503\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.858227 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zpz\" (UniqueName: \"kubernetes.io/projected/bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5-kube-api-access-z7zpz\") pod \"openshift-config-operator-7777fb866f-6mbk6\" (UID: \"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.875666 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n5v4\" (UniqueName: \"kubernetes.io/projected/6cdca78e-93e0-4e36-94a6-51a65c42ef75-kube-api-access-7n5v4\") pod \"etcd-operator-b45778765-xszsl\" (UID: \"6cdca78e-93e0-4e36-94a6-51a65c42ef75\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.885722 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.898890 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.904402 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tmd4d\" (UID: \"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.913135 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.938659 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rklvs"] Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.939333 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.941949 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9hhs\" (UniqueName: \"kubernetes.io/projected/9b4e607c-15af-4ba5-8501-eb50c38c522d-kube-api-access-p9hhs\") pod \"dns-operator-744455d44c-smwgx\" (UID: \"9b4e607c-15af-4ba5-8501-eb50c38c522d\") " pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.951997 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jpnn\" (UniqueName: \"kubernetes.io/projected/7ca8ccaf-fa72-4f5d-ba14-721c90583efb-kube-api-access-4jpnn\") pod \"router-default-5444994796-vb728\" (UID: \"7ca8ccaf-fa72-4f5d-ba14-721c90583efb\") " pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.982492 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/919c90bd-0d06-4af1-89b2-b3d3aec78ad4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8mhbv\" (UID: \"919c90bd-0d06-4af1-89b2-b3d3aec78ad4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.983868 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" Oct 01 07:08:04 crc kubenswrapper[4837]: I1001 07:08:04.984174 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.001780 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.009395 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwljd\" (UniqueName: \"kubernetes.io/projected/065e771b-41ef-4984-aacb-6dd2e1231815-kube-api-access-fwljd\") pod \"openshift-controller-manager-operator-756b6f6bc6-9txxm\" (UID: \"065e771b-41ef-4984-aacb-6dd2e1231815\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.009615 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wkd2\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-kube-api-access-5wkd2\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.015739 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/47c7934b-3c05-4fb1-aef3-36f6f2b76d1f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z89gn\" (UID: \"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.037008 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87ac3d3f-dc65-4e67-8a79-76d9fbcdece4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zsd2v\" (UID: \"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.065138 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbp4r\" (UniqueName: \"kubernetes.io/projected/873e7d6c-b8c8-4b7a-a694-d1a6ae624d95-kube-api-access-dbp4r\") pod \"kube-storage-version-migrator-operator-b67b599dd-7qrjz\" (UID: \"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087092 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087382 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087412 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/880f40fa-c9a4-4fff-a358-4750e9788839-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087436 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087486 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkq2b\" (UniqueName: \"kubernetes.io/projected/880f40fa-c9a4-4fff-a358-4750e9788839-kube-api-access-fkq2b\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087510 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087532 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtzng\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087561 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087590 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087619 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.087654 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/880f40fa-c9a4-4fff-a358-4750e9788839-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.088048 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.58803549 +0000 UTC m=+142.429642935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.193234 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.194955 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.694916139 +0000 UTC m=+142.536523594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195051 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k52ms\" (UniqueName: \"kubernetes.io/projected/15fc768b-5e64-4f99-be37-4d7f33158487-kube-api-access-k52ms\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195196 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/880f40fa-c9a4-4fff-a358-4750e9788839-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195283 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195314 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195345 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/880f40fa-c9a4-4fff-a358-4750e9788839-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195371 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195435 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/15fc768b-5e64-4f99-be37-4d7f33158487-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195468 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkq2b\" (UniqueName: \"kubernetes.io/projected/880f40fa-c9a4-4fff-a358-4750e9788839-kube-api-access-fkq2b\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195510 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195549 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtzng\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195594 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.195613 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.196936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/880f40fa-c9a4-4fff-a358-4750e9788839-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.198016 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.198789 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.698766791 +0000 UTC m=+142.540374246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.198930 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.200597 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.212137 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.212798 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/880f40fa-c9a4-4fff-a358-4750e9788839-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.218078 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-822f9"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.244642 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.245745 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.250893 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.261228 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.261436 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkq2b\" (UniqueName: \"kubernetes.io/projected/880f40fa-c9a4-4fff-a358-4750e9788839-kube-api-access-fkq2b\") pod \"openshift-apiserver-operator-796bbdcf4f-6q4c5\" (UID: \"880f40fa-c9a4-4fff-a358-4750e9788839\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.268559 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.272999 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtzng\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.278662 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.283183 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.292769 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.299359 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.300455 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.800431554 +0000 UTC m=+142.642039009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.300945 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-registration-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.300988 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-cabundle\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301149 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301181 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301197 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51f77b47-e87f-457c-9f20-b61f5f3ea339-proxy-tls\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301213 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-mountpoint-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301230 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-webhook-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301266 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-node-bootstrap-token\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301299 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-key\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.301322 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/363b7780-466e-4cff-bee5-8ee51d41d931-config-volume\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303733 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqqt\" (UniqueName: \"kubernetes.io/projected/51f77b47-e87f-457c-9f20-b61f5f3ea339-kube-api-access-zsqqt\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303776 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303792 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303866 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-csi-data-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303884 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwthj\" (UniqueName: \"kubernetes.io/projected/6e0a8f48-6289-4c87-aba6-95ae84b691d3-kube-api-access-rwthj\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303899 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk8gk\" (UniqueName: \"kubernetes.io/projected/5858f224-e8b1-4266-b0ab-1d5e72c884c9-kube-api-access-wk8gk\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.303954 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304007 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6plm\" (UniqueName: \"kubernetes.io/projected/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-kube-api-access-p6plm\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304022 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ftk\" (UniqueName: \"kubernetes.io/projected/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-kube-api-access-r4ftk\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304047 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phzzc\" (UniqueName: \"kubernetes.io/projected/363b7780-466e-4cff-bee5-8ee51d41d931-kube-api-access-phzzc\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304062 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbxqt\" (UniqueName: \"kubernetes.io/projected/851e1990-309d-43f3-ab4f-6d18d4f61ebd-kube-api-access-nbxqt\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304076 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-apiservice-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304102 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304146 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k52ms\" (UniqueName: \"kubernetes.io/projected/15fc768b-5e64-4f99-be37-4d7f33158487-kube-api-access-k52ms\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e0a8f48-6289-4c87-aba6-95ae84b691d3-proxy-tls\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304251 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8k5t\" (UniqueName: \"kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.304275 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b8012dc-8975-4a9d-86c3-436e7fca87fa-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305060 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nswm\" (UniqueName: \"kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305137 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/363b7780-466e-4cff-bee5-8ee51d41d931-metrics-tls\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305178 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-certs\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305202 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x965f\" (UniqueName: \"kubernetes.io/projected/0ff14da2-35ae-43a1-833f-f79ffbce7f57-kube-api-access-x965f\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305223 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-srv-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305256 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63149934-868e-4a3c-996b-aa7fcdd4cdac-tmpfs\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305436 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305518 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrwdx\" (UniqueName: \"kubernetes.io/projected/1985d236-fa0b-441d-824a-82c8be177844-kube-api-access-wrwdx\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305649 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd4rf\" (UniqueName: \"kubernetes.io/projected/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-kube-api-access-jd4rf\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305814 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-plugins-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305843 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e0a8f48-6289-4c87-aba6-95ae84b691d3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305910 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzdrz\" (UniqueName: \"kubernetes.io/projected/63149934-868e-4a3c-996b-aa7fcdd4cdac-kube-api-access-pzdrz\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff14da2-35ae-43a1-833f-f79ffbce7f57-serving-cert\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305970 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.305987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-socket-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306033 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff14da2-35ae-43a1-833f-f79ffbce7f57-config\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306060 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4szv5\" (UniqueName: \"kubernetes.io/projected/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-kube-api-access-4szv5\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306087 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-images\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306650 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/15fc768b-5e64-4f99-be37-4d7f33158487-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306674 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1985d236-fa0b-441d-824a-82c8be177844-cert\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306699 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-srv-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306718 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcsb6\" (UniqueName: \"kubernetes.io/projected/b281bf11-93c6-4cc2-9f60-9ffe709d7689-kube-api-access-kcsb6\") pod \"migrator-59844c95c7-n5kl8\" (UID: \"b281bf11-93c6-4cc2-9f60-9ffe709d7689\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.306735 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlk9m\" (UniqueName: \"kubernetes.io/projected/7b8012dc-8975-4a9d-86c3-436e7fca87fa-kube-api-access-qlk9m\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.309147 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.809123981 +0000 UTC m=+142.650731546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.314974 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/15fc768b-5e64-4f99-be37-4d7f33158487-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: W1001 07:08:05.338728 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25fe367b_044a_4d96_9658_eaad1a16582a.slice/crio-fa4dd93a848f180c7a538f89b97892bd2fe001d7fe603371059da95437f4fee9 WatchSource:0}: Error finding container fa4dd93a848f180c7a538f89b97892bd2fe001d7fe603371059da95437f4fee9: Status 404 returned error can't find the container with id fa4dd93a848f180c7a538f89b97892bd2fe001d7fe603371059da95437f4fee9 Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.357050 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k52ms\" (UniqueName: \"kubernetes.io/projected/15fc768b-5e64-4f99-be37-4d7f33158487-kube-api-access-k52ms\") pod \"package-server-manager-789f6589d5-sl6hv\" (UID: \"15fc768b-5e64-4f99-be37-4d7f33158487\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.407416 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.407747 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.907727653 +0000 UTC m=+142.749335108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408649 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408700 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408724 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51f77b47-e87f-457c-9f20-b61f5f3ea339-proxy-tls\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408745 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-mountpoint-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408770 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-webhook-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408794 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-node-bootstrap-token\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.408816 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-key\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.409094 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/363b7780-466e-4cff-bee5-8ee51d41d931-config-volume\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.409187 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-mountpoint-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.416972 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-profile-collector-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.419119 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420077 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqqt\" (UniqueName: \"kubernetes.io/projected/51f77b47-e87f-457c-9f20-b61f5f3ea339-kube-api-access-zsqqt\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420123 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420144 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420185 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-csi-data-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420207 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwthj\" (UniqueName: \"kubernetes.io/projected/6e0a8f48-6289-4c87-aba6-95ae84b691d3-kube-api-access-rwthj\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420230 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420267 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk8gk\" (UniqueName: \"kubernetes.io/projected/5858f224-e8b1-4266-b0ab-1d5e72c884c9-kube-api-access-wk8gk\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420290 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420366 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6plm\" (UniqueName: \"kubernetes.io/projected/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-kube-api-access-p6plm\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420386 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ftk\" (UniqueName: \"kubernetes.io/projected/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-kube-api-access-r4ftk\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420446 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phzzc\" (UniqueName: \"kubernetes.io/projected/363b7780-466e-4cff-bee5-8ee51d41d931-kube-api-access-phzzc\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420466 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbxqt\" (UniqueName: \"kubernetes.io/projected/851e1990-309d-43f3-ab4f-6d18d4f61ebd-kube-api-access-nbxqt\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420486 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-apiservice-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420525 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420595 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8k5t\" (UniqueName: \"kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420617 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b8012dc-8975-4a9d-86c3-436e7fca87fa-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420643 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e0a8f48-6289-4c87-aba6-95ae84b691d3-proxy-tls\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420736 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nswm\" (UniqueName: \"kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420755 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/363b7780-466e-4cff-bee5-8ee51d41d931-metrics-tls\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.420793 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-certs\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.422133 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x965f\" (UniqueName: \"kubernetes.io/projected/0ff14da2-35ae-43a1-833f-f79ffbce7f57-kube-api-access-x965f\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.422160 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-srv-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423314 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63149934-868e-4a3c-996b-aa7fcdd4cdac-tmpfs\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423344 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrwdx\" (UniqueName: \"kubernetes.io/projected/1985d236-fa0b-441d-824a-82c8be177844-kube-api-access-wrwdx\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423378 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423395 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd4rf\" (UniqueName: \"kubernetes.io/projected/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-kube-api-access-jd4rf\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423417 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-plugins-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423471 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e0a8f48-6289-4c87-aba6-95ae84b691d3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.423606 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424667 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzdrz\" (UniqueName: \"kubernetes.io/projected/63149934-868e-4a3c-996b-aa7fcdd4cdac-kube-api-access-pzdrz\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424748 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff14da2-35ae-43a1-833f-f79ffbce7f57-serving-cert\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424804 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424824 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-socket-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424840 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff14da2-35ae-43a1-833f-f79ffbce7f57-config\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424875 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4szv5\" (UniqueName: \"kubernetes.io/projected/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-kube-api-access-4szv5\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424911 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-images\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424930 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1985d236-fa0b-441d-824a-82c8be177844-cert\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424967 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcsb6\" (UniqueName: \"kubernetes.io/projected/b281bf11-93c6-4cc2-9f60-9ffe709d7689-kube-api-access-kcsb6\") pod \"migrator-59844c95c7-n5kl8\" (UID: \"b281bf11-93c6-4cc2-9f60-9ffe709d7689\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.424985 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlk9m\" (UniqueName: \"kubernetes.io/projected/7b8012dc-8975-4a9d-86c3-436e7fca87fa-kube-api-access-qlk9m\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.425002 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-srv-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.425051 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-registration-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.425073 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-cabundle\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.426092 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-cabundle\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.427238 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/63149934-868e-4a3c-996b-aa7fcdd4cdac-tmpfs\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.427725 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.427775 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.428029 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ff14da2-35ae-43a1-833f-f79ffbce7f57-config\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.428919 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/51f77b47-e87f-457c-9f20-b61f5f3ea339-images\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.429026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-plugins-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.421075 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/363b7780-466e-4cff-bee5-8ee51d41d931-config-volume\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.429776 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:05.929762799 +0000 UTC m=+142.771370254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.431921 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-csi-data-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.432253 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-socket-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.432303 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-registration-dir\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.433173 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e0a8f48-6289-4c87-aba6-95ae84b691d3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.433204 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.438405 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51f77b47-e87f-457c-9f20-b61f5f3ea339-proxy-tls\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.439138 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.439496 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5858f224-e8b1-4266-b0ab-1d5e72c884c9-srv-cert\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.443673 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851e1990-309d-43f3-ab4f-6d18d4f61ebd-signing-key\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.444592 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/363b7780-466e-4cff-bee5-8ee51d41d931-metrics-tls\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.444894 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6e0a8f48-6289-4c87-aba6-95ae84b691d3-proxy-tls\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.444976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1985d236-fa0b-441d-824a-82c8be177844-cert\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.445160 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-node-bootstrap-token\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.445248 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-srv-cert\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.445848 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-webhook-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.448138 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.448199 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-certs\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.448304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.448563 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b8012dc-8975-4a9d-86c3-436e7fca87fa-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.449477 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63149934-868e-4a3c-996b-aa7fcdd4cdac-apiservice-cert\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.453467 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.455194 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ff14da2-35ae-43a1-833f-f79ffbce7f57-serving-cert\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.459260 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ftk\" (UniqueName: \"kubernetes.io/projected/1b79aedb-c9cb-46b7-8e26-6b97b8d017af-kube-api-access-r4ftk\") pod \"catalog-operator-68c6474976-mmx44\" (UID: \"1b79aedb-c9cb-46b7-8e26-6b97b8d017af\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.475657 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.478629 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqqt\" (UniqueName: \"kubernetes.io/projected/51f77b47-e87f-457c-9f20-b61f5f3ea339-kube-api-access-zsqqt\") pod \"machine-config-operator-74547568cd-ggxrb\" (UID: \"51f77b47-e87f-457c-9f20-b61f5f3ea339\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.516344 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phzzc\" (UniqueName: \"kubernetes.io/projected/363b7780-466e-4cff-bee5-8ee51d41d931-kube-api-access-phzzc\") pod \"dns-default-m78s5\" (UID: \"363b7780-466e-4cff-bee5-8ee51d41d931\") " pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.521296 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbxqt\" (UniqueName: \"kubernetes.io/projected/851e1990-309d-43f3-ab4f-6d18d4f61ebd-kube-api-access-nbxqt\") pod \"service-ca-9c57cc56f-rwwps\" (UID: \"851e1990-309d-43f3-ab4f-6d18d4f61ebd\") " pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.529480 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.530220 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.030193973 +0000 UTC m=+142.871801428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.579578 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrwdx\" (UniqueName: \"kubernetes.io/projected/1985d236-fa0b-441d-824a-82c8be177844-kube-api-access-wrwdx\") pod \"ingress-canary-45m2c\" (UID: \"1985d236-fa0b-441d-824a-82c8be177844\") " pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.585572 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8k5t\" (UniqueName: \"kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t\") pod \"marketplace-operator-79b997595-j2drz\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.598591 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4szv5\" (UniqueName: \"kubernetes.io/projected/1ba7ee8e-d52c-403d-af55-25abe87c8c3c-kube-api-access-4szv5\") pod \"multus-admission-controller-857f4d67dd-rlk6w\" (UID: \"1ba7ee8e-d52c-403d-af55-25abe87c8c3c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.616909 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.633604 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.634003 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.133987947 +0000 UTC m=+142.975595403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.635214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nswm\" (UniqueName: \"kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm\") pod \"collect-profiles-29321700-vhv2l\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.639113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk8gk\" (UniqueName: \"kubernetes.io/projected/5858f224-e8b1-4266-b0ab-1d5e72c884c9-kube-api-access-wk8gk\") pod \"olm-operator-6b444d44fb-nqcvx\" (UID: \"5858f224-e8b1-4266-b0ab-1d5e72c884c9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.654759 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.657304 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.657612 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.665221 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-25t4p"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.666956 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cn9dw"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.668922 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.671382 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.675037 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzdrz\" (UniqueName: \"kubernetes.io/projected/63149934-868e-4a3c-996b-aa7fcdd4cdac-kube-api-access-pzdrz\") pod \"packageserver-d55dfcdfc-2bc7z\" (UID: \"63149934-868e-4a3c-996b-aa7fcdd4cdac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.679399 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd4rf\" (UniqueName: \"kubernetes.io/projected/dfef3acd-6b92-43ee-a0fe-615a0d62ee14-kube-api-access-jd4rf\") pod \"csi-hostpathplugin-9lclf\" (UID: \"dfef3acd-6b92-43ee-a0fe-615a0d62ee14\") " pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.685761 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.698409 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:05 crc kubenswrapper[4837]: W1001 07:08:05.702459 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod485c8ce3_3b62_4fc6_9a57_92b3c3afc4d7.slice/crio-2ac94c748bf6683c1e3f5760d7e297ebc5cee9caf7ca98462a5c4a8ae5d44573 WatchSource:0}: Error finding container 2ac94c748bf6683c1e3f5760d7e297ebc5cee9caf7ca98462a5c4a8ae5d44573: Status 404 returned error can't find the container with id 2ac94c748bf6683c1e3f5760d7e297ebc5cee9caf7ca98462a5c4a8ae5d44573 Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.706241 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwthj\" (UniqueName: \"kubernetes.io/projected/6e0a8f48-6289-4c87-aba6-95ae84b691d3-kube-api-access-rwthj\") pod \"machine-config-controller-84d6567774-z5w2r\" (UID: \"6e0a8f48-6289-4c87-aba6-95ae84b691d3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.727627 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlk9m\" (UniqueName: \"kubernetes.io/projected/7b8012dc-8975-4a9d-86c3-436e7fca87fa-kube-api-access-qlk9m\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vht5\" (UID: \"7b8012dc-8975-4a9d-86c3-436e7fca87fa\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.739285 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.739679 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.739981 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.740355 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.240338533 +0000 UTC m=+143.081945988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.746510 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-45m2c" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.747896 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcsb6\" (UniqueName: \"kubernetes.io/projected/b281bf11-93c6-4cc2-9f60-9ffe709d7689-kube-api-access-kcsb6\") pod \"migrator-59844c95c7-n5kl8\" (UID: \"b281bf11-93c6-4cc2-9f60-9ffe709d7689\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.755124 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6plm\" (UniqueName: \"kubernetes.io/projected/6ab1bcc5-281a-4315-81cd-4a803eccfe6f-kube-api-access-p6plm\") pod \"machine-config-server-hjphq\" (UID: \"6ab1bcc5-281a-4315-81cd-4a803eccfe6f\") " pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.770335 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" event={"ID":"6042d50d-4f20-47d8-a9d8-74c8424aff37","Type":"ContainerStarted","Data":"52265347b20baea6d5ef4077ab7fdebe50599ffa8dcd6321298420074d19d8e3"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.779677 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vb728" event={"ID":"7ca8ccaf-fa72-4f5d-ba14-721c90583efb","Type":"ContainerStarted","Data":"4ce3b30135b5eff61a8af3ad38ede89b31df12b460ebbec16c1bf723ecb03297"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.779741 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vb728" event={"ID":"7ca8ccaf-fa72-4f5d-ba14-721c90583efb","Type":"ContainerStarted","Data":"8ae9497300c7e818b07f93e2e435fb2a5002f1383b3763f1f6c3ea5212a938e3"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.789023 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" event={"ID":"25fe367b-044a-4d96-9658-eaad1a16582a","Type":"ContainerStarted","Data":"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.789072 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" event={"ID":"25fe367b-044a-4d96-9658-eaad1a16582a","Type":"ContainerStarted","Data":"fa4dd93a848f180c7a538f89b97892bd2fe001d7fe603371059da95437f4fee9"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.790034 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.791199 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" event={"ID":"db418233-a22c-4134-8414-93c3bbb3bf1f","Type":"ContainerStarted","Data":"9fea9b0b35632bfe192a2e23303465dc0b0b320f40f343981fc8ffad2db8c460"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.791224 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" event={"ID":"db418233-a22c-4134-8414-93c3bbb3bf1f","Type":"ContainerStarted","Data":"e0e7a47b9ade2095e81263b98da03c96d8612b168651d5da02fd19df5280e8d8"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.791928 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mg4h6" event={"ID":"c5794771-53a6-46dd-bd10-b37f6534b87b","Type":"ContainerStarted","Data":"87750c41e5bf324c1aebdfaa153201e1b15e7a0f4b66af6f2922128d7396d607"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.793825 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" event={"ID":"67a8d1d3-d463-4e79-be09-1e552f2fde05","Type":"ContainerStarted","Data":"4cf40f4fa7d469b1056e085915288e22feb55d349e236116686e9f727501ece7"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.793849 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" event={"ID":"67a8d1d3-d463-4e79-be09-1e552f2fde05","Type":"ContainerStarted","Data":"563ac7625019c9b4688a3097366c2f90fc351d1bd602257672dbe8471a2a28ba"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.793862 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" event={"ID":"67a8d1d3-d463-4e79-be09-1e552f2fde05","Type":"ContainerStarted","Data":"f152232d18c7659c5c89dd41c7fa27ef61ece5f54c7e48eea1fb1828e9e44924"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.795089 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" event={"ID":"2abd0aad-2266-4b79-9eb0-f03154ed176c","Type":"ContainerStarted","Data":"36e75c1e88c2e87ec696bb6a9bd7d0f583f3d8e8e2fafc5c8ea20d29db80b682"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.795122 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" event={"ID":"2abd0aad-2266-4b79-9eb0-f03154ed176c","Type":"ContainerStarted","Data":"9490f80fee3d0e2bbad6c9c1edfba8d0595871f5e0d4c695fe11fc1b8129c1a8"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.799598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x965f\" (UniqueName: \"kubernetes.io/projected/0ff14da2-35ae-43a1-833f-f79ffbce7f57-kube-api-access-x965f\") pod \"service-ca-operator-777779d784-4fsxp\" (UID: \"0ff14da2-35ae-43a1-833f-f79ffbce7f57\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.806044 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xszsl"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.814882 4837 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4cx5v container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.814949 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.837962 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.838001 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" event={"ID":"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3","Type":"ContainerStarted","Data":"96e4ce3b1dc9f8614798ba9dfb5d20ac11992a7823b66ec2e8cff9ca85772d9a"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.838041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" event={"ID":"968df061-3925-4301-9444-1b02e679de29","Type":"ContainerStarted","Data":"937595e2aae834fdfece95ab2232284bcf23c32af2fbaeb2efff8311c84a30b8"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.838053 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" event={"ID":"968df061-3925-4301-9444-1b02e679de29","Type":"ContainerStarted","Data":"9e4dcf6d83c6d11c07805b28e63a1272cc5ee0e27b032371fd5ac5146f8d1af5"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.838541 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" event={"ID":"a6fb593e-7716-401b-b74c-de0145159433","Type":"ContainerStarted","Data":"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.838564 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" event={"ID":"a6fb593e-7716-401b-b74c-de0145159433","Type":"ContainerStarted","Data":"5c5b130dcdd21472bcf5d14db54a7b11884bc9cc0c88c441b9cf3856ab288ffe"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.839273 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.842221 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.846278 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" event={"ID":"d5baef47-d279-4a19-88f6-271f843d6594","Type":"ContainerStarted","Data":"d0c66f5ee9dc66ee825d12370457a2b004aacf7397dc03b37ca2bf047c82e708"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.846350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" event={"ID":"d5baef47-d279-4a19-88f6-271f843d6594","Type":"ContainerStarted","Data":"84f0c84eeb625e0d1aefd179ecdf2a3ea05fac708d9faca5ff9d94ff54ca1ba1"} Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.849748 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.851314 4837 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mlw9m container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.851357 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" podUID="a6fb593e-7716-401b-b74c-de0145159433" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.851463 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.351431003 +0000 UTC m=+143.193038658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.882329 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.910116 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.911266 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.932874 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.933046 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.934430 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.943302 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.943715 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.443677943 +0000 UTC m=+143.285285398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.943826 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.946933 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v"] Oct 01 07:08:05 crc kubenswrapper[4837]: E1001 07:08:05.947592 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.447576785 +0000 UTC m=+143.289184240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.959462 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-smwgx"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.959742 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d"] Oct 01 07:08:05 crc kubenswrapper[4837]: I1001 07:08:05.982405 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.002790 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.004774 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" Oct 01 07:08:06 crc kubenswrapper[4837]: W1001 07:08:06.021013 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f6c496c_0b1e_4e04_a70c_56ae687b3503.slice/crio-76e9766efff55d77afee6ddebd18e9f2a0112e2218b198ade92574f7bf0383e7 WatchSource:0}: Error finding container 76e9766efff55d77afee6ddebd18e9f2a0112e2218b198ade92574f7bf0383e7: Status 404 returned error can't find the container with id 76e9766efff55d77afee6ddebd18e9f2a0112e2218b198ade92574f7bf0383e7 Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.021805 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hjphq" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.029563 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.037295 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:06 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:06 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:06 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.037353 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.054831 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.554804082 +0000 UTC m=+143.396411537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.057664 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.057913 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.058496 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.55848268 +0000 UTC m=+143.400090135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.086902 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.132042 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.171907 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.172550 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.672506628 +0000 UTC m=+143.514114233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.189290 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.277277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.278111 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.778088946 +0000 UTC m=+143.619696401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: W1001 07:08:06.308758 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15fc768b_5e64_4f99_be37_4d7f33158487.slice/crio-63faf63ce09bccc7452ab43561503baeb68dd5e9790e20e67043c4fdcfc9d795 WatchSource:0}: Error finding container 63faf63ce09bccc7452ab43561503baeb68dd5e9790e20e67043c4fdcfc9d795: Status 404 returned error can't find the container with id 63faf63ce09bccc7452ab43561503baeb68dd5e9790e20e67043c4fdcfc9d795 Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.333769 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-45m2c"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.379017 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.379522 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.879501954 +0000 UTC m=+143.721109409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.405192 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mkxdr" podStartSLOduration=122.405165686 podStartE2EDuration="2m2.405165686s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:06.378239504 +0000 UTC m=+143.219846959" watchObservedRunningTime="2025-10-01 07:08:06.405165686 +0000 UTC m=+143.246773141" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.406321 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.456853 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9pf94" podStartSLOduration=121.456828488 podStartE2EDuration="2m1.456828488s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:06.454537473 +0000 UTC m=+143.296144928" watchObservedRunningTime="2025-10-01 07:08:06.456828488 +0000 UTC m=+143.298435933" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.482318 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.482957 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:06.982930031 +0000 UTC m=+143.824537676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.518593 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.533515 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.587101 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.587292 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.087257928 +0000 UTC m=+143.928865383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.588010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.588423 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.088407585 +0000 UTC m=+143.930015040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.666612 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" podStartSLOduration=122.66658496 podStartE2EDuration="2m2.66658496s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:06.666503258 +0000 UTC m=+143.508110723" watchObservedRunningTime="2025-10-01 07:08:06.66658496 +0000 UTC m=+143.508192415" Oct 01 07:08:06 crc kubenswrapper[4837]: W1001 07:08:06.682124 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7fc4d5e_2b1f_4db6_8e0c_d261dc3df117.slice/crio-e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0 WatchSource:0}: Error finding container e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0: Status 404 returned error can't find the container with id e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0 Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.692450 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.692772 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.192743904 +0000 UTC m=+144.034351359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.692936 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.693388 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.193381568 +0000 UTC m=+144.034989023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.713032 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rlk6w"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.714910 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.749085 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m78s5"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.821865 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.822258 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.322231611 +0000 UTC m=+144.163839066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.823029 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.823865 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.323849409 +0000 UTC m=+144.165456864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.837013 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.854098 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9lclf"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.874878 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mg4h6" event={"ID":"c5794771-53a6-46dd-bd10-b37f6534b87b","Type":"ContainerStarted","Data":"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.908478 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hjphq" event={"ID":"6ab1bcc5-281a-4315-81cd-4a803eccfe6f","Type":"ContainerStarted","Data":"be1aad8cc74a709aa7156295d44488b5df4deabe9c4e6e2289251b75d9aa2d9a"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.911408 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rwwps"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.917559 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" event={"ID":"1b79aedb-c9cb-46b7-8e26-6b97b8d017af","Type":"ContainerStarted","Data":"a45dc78c746e4558ceb388d67419ff66546b50a3a6e03a4c1603dafc330bf357"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.919937 4837 generic.go:334] "Generic (PLEG): container finished" podID="785b6d63-389e-4eb8-ba3b-67f23d5dd3c3" containerID="b161c9cf1fbb219aaeb842f8bfb0c5cc5c0f6c9d59a36fb430a6039293b991b6" exitCode=0 Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.920023 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" event={"ID":"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3","Type":"ContainerDied","Data":"b161c9cf1fbb219aaeb842f8bfb0c5cc5c0f6c9d59a36fb430a6039293b991b6"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.923409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" event={"ID":"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f","Type":"ContainerStarted","Data":"a65f82d0e904eb9dda5fcfb82c421bfeb51a3d207db2ef1e6a5222a5e31adb32"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.924040 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:06 crc kubenswrapper[4837]: E1001 07:08:06.925295 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.425276958 +0000 UTC m=+144.266884413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.927344 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" event={"ID":"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c","Type":"ContainerStarted","Data":"2141cbc35941744dfd0b3ba3441001743fd271a9664a9ef716cdacb1c88d1b08"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.942204 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" event={"ID":"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7","Type":"ContainerStarted","Data":"e93e75c3fdb3e67d187cb17ab0e1a3224cf7a0fdb939aa6f2339f12206cf129a"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.942287 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" event={"ID":"485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7","Type":"ContainerStarted","Data":"2ac94c748bf6683c1e3f5760d7e297ebc5cee9caf7ca98462a5c4a8ae5d44573"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.943444 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.952356 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" event={"ID":"6042d50d-4f20-47d8-a9d8-74c8424aff37","Type":"ContainerStarted","Data":"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.953578 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.962035 4837 patch_prober.go:28] interesting pod/console-operator-58897d9998-cn9dw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.962088 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" podUID="485c8ce3-3b62-4fc6-9a57-92b3c3afc4d7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.963526 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.965381 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.971124 4837 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-t8vmh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.971175 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.971377 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" event={"ID":"9b4e607c-15af-4ba5-8501-eb50c38c522d","Type":"ContainerStarted","Data":"219b4094be02966e5898fd330061d0831d97631d8cfac25785d0bfd49120ba7b"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.974429 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" event={"ID":"15fc768b-5e64-4f99-be37-4d7f33158487","Type":"ContainerStarted","Data":"63faf63ce09bccc7452ab43561503baeb68dd5e9790e20e67043c4fdcfc9d795"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.984472 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5"] Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.986946 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" event={"ID":"065e771b-41ef-4984-aacb-6dd2e1231815","Type":"ContainerStarted","Data":"58491a3a71510d50a47dc77784cd0a387cb5ff05bdaf19d532b6b88d1ed82e76"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.990277 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" event={"ID":"880f40fa-c9a4-4fff-a358-4750e9788839","Type":"ContainerStarted","Data":"0b6f378245f289cfbede9561eb1cb7cfea8db6841f9a7261d6d0e1b335f74cbc"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.996948 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" event={"ID":"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95","Type":"ContainerStarted","Data":"b7d181f946acf6483c5bb9770280d949b08f4a51bb4091cbf06f6b301d78667a"} Oct 01 07:08:06 crc kubenswrapper[4837]: I1001 07:08:06.997052 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" event={"ID":"873e7d6c-b8c8-4b7a-a694-d1a6ae624d95","Type":"ContainerStarted","Data":"b6d06d8d1146fa1a2be34504a91c2a5fc21ba293dc9efceb1cb632de1157d1e0"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.000733 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" event={"ID":"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5","Type":"ContainerStarted","Data":"193459dcb43c9d9915ef1c82f0b6e379cb4dd361935d1ecefaff70211ab5dc15"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.004328 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" event={"ID":"4f6c496c-0b1e-4e04-a70c-56ae687b3503","Type":"ContainerStarted","Data":"76e9766efff55d77afee6ddebd18e9f2a0112e2218b198ade92574f7bf0383e7"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.009582 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:07 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:07 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:07 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.010021 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.027361 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.029082 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.529060503 +0000 UTC m=+144.370667958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: W1001 07:08:07.035467 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ba7ee8e_d52c_403d_af55_25abe87c8c3c.slice/crio-e40cb9e4128e4099fcea1d79526d9e2517ae74bbf0d42a6d2896cfb9e8c739ff WatchSource:0}: Error finding container e40cb9e4128e4099fcea1d79526d9e2517ae74bbf0d42a6d2896cfb9e8c739ff: Status 404 returned error can't find the container with id e40cb9e4128e4099fcea1d79526d9e2517ae74bbf0d42a6d2896cfb9e8c739ff Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.041382 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" event={"ID":"d5baef47-d279-4a19-88f6-271f843d6594","Type":"ContainerStarted","Data":"570ab518bb08be8ced435d1440e028212a44c156f3957f356d0c43ace9db1339"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.043292 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" event={"ID":"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117","Type":"ContainerStarted","Data":"e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.044919 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" event={"ID":"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4","Type":"ContainerStarted","Data":"280b98745a4e1b5fee0234b4d509eca1c5fc5e75fe29615475aa22ea9e01f76a"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.052412 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" event={"ID":"db418233-a22c-4134-8414-93c3bbb3bf1f","Type":"ContainerStarted","Data":"48b602ca6d546592cb70f1e382a03d7526b5a0320499b118ca3fb924d1417593"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.054315 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" event={"ID":"6cdca78e-93e0-4e36-94a6-51a65c42ef75","Type":"ContainerStarted","Data":"8e42fa70432ea1c6f813d3653c17ad3331eb768ce7f93afda0f127ff274c674f"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.072970 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" event={"ID":"f4ae7b9d-d7f0-4367-860f-81106f4b9d63","Type":"ContainerStarted","Data":"ed61df8eab6b90f957df9df68cda092c623b7d42af4ad0b4e092ff30ea7103a9"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.074079 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-25t4p" event={"ID":"5be4a157-1e4f-4e93-8f85-2c01e1f60ff1","Type":"ContainerStarted","Data":"3ac90a4f67116270ec0bf1e6c059d654ebb5ca27eefc5df9571e5d1ddbb4dad3"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.075299 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" event={"ID":"919c90bd-0d06-4af1-89b2-b3d3aec78ad4","Type":"ContainerStarted","Data":"bb86b92844fefa1c4b8ef988fc1636b9267cd4a0cdd738630566dd9f13fbfe21"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.116634 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-45m2c" event={"ID":"1985d236-fa0b-441d-824a-82c8be177844","Type":"ContainerStarted","Data":"245b792c7568338ca09722843dedd98944ea42cf0eb8451031f5c7ecb2f3a291"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.119305 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" event={"ID":"5858f224-e8b1-4266-b0ab-1d5e72c884c9","Type":"ContainerStarted","Data":"98a38e4ec7aecfe4c1c81f03d7e6c43389a6fc20c28fc6662654ad1e9b277714"} Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.123050 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-vb728" podStartSLOduration=122.123004793 podStartE2EDuration="2m2.123004793s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.121066496 +0000 UTC m=+143.962673951" watchObservedRunningTime="2025-10-01 07:08:07.123004793 +0000 UTC m=+143.964612248" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.127978 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.128233 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.628202927 +0000 UTC m=+144.469810382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.128357 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.129519 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.629486537 +0000 UTC m=+144.471094192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.147791 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp"] Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.210045 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tx9ss" podStartSLOduration=123.210016557 podStartE2EDuration="2m3.210016557s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.179969611 +0000 UTC m=+144.021577076" watchObservedRunningTime="2025-10-01 07:08:07.210016557 +0000 UTC m=+144.051624012" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.222951 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.230005 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.232078 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.732055393 +0000 UTC m=+144.573662848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.297332 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" podStartSLOduration=123.297302108 podStartE2EDuration="2m3.297302108s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.295194658 +0000 UTC m=+144.136802113" watchObservedRunningTime="2025-10-01 07:08:07.297302108 +0000 UTC m=+144.138909563" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.316932 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.332162 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.332537 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.832523418 +0000 UTC m=+144.674130873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.433714 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.434073 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:07.934055259 +0000 UTC m=+144.775662714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.474768 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-822f9" podStartSLOduration=122.474670218 podStartE2EDuration="2m2.474670218s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.437339697 +0000 UTC m=+144.278947152" watchObservedRunningTime="2025-10-01 07:08:07.474670218 +0000 UTC m=+144.316277683" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.535050 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.535740 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.035726734 +0000 UTC m=+144.877334189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.606475 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mdg4v" podStartSLOduration=123.60645369 podStartE2EDuration="2m3.60645369s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.605552088 +0000 UTC m=+144.447159553" watchObservedRunningTime="2025-10-01 07:08:07.60645369 +0000 UTC m=+144.448061145" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.636630 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.637047 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.137030449 +0000 UTC m=+144.978637894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.738003 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.738404 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.238387285 +0000 UTC m=+145.079994740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.772834 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mg4h6" podStartSLOduration=123.772799276 podStartE2EDuration="2m3.772799276s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.764500518 +0000 UTC m=+144.606107983" watchObservedRunningTime="2025-10-01 07:08:07.772799276 +0000 UTC m=+144.614406731" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.820069 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" podStartSLOduration=123.820044803 podStartE2EDuration="2m3.820044803s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:07.818487536 +0000 UTC m=+144.660094991" watchObservedRunningTime="2025-10-01 07:08:07.820044803 +0000 UTC m=+144.661652258" Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.840652 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.841938 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.341917694 +0000 UTC m=+145.183525149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:07 crc kubenswrapper[4837]: I1001 07:08:07.944574 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:07 crc kubenswrapper[4837]: E1001 07:08:07.945540 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.445499954 +0000 UTC m=+145.287107589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.008440 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:08 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:08 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:08 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.008548 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.046465 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.047150 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.547124307 +0000 UTC m=+145.388731762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.112221 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" podStartSLOduration=123.112199789 podStartE2EDuration="2m3.112199789s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.077176734 +0000 UTC m=+144.918784179" watchObservedRunningTime="2025-10-01 07:08:08.112199789 +0000 UTC m=+144.953807244" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.149242 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.149658 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.649640902 +0000 UTC m=+145.491248357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.152328 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" event={"ID":"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3","Type":"ContainerStarted","Data":"8dc0f1bf3784db0ea1de4e44e452410c75316546ea1c1142fe939d6e357d6f32"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.183239 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-25t4p" event={"ID":"5be4a157-1e4f-4e93-8f85-2c01e1f60ff1","Type":"ContainerStarted","Data":"64d710f92c98782ff90ac1c22d5e7259e77c78bbe7cb3f4412a2a8c32ff1870a"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.184459 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.202898 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.202982 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.234038 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" event={"ID":"5858f224-e8b1-4266-b0ab-1d5e72c884c9","Type":"ContainerStarted","Data":"404554c5d611e0b73029526b90863860548273acdad7c2e2173d9eade014887a"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.235757 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.245172 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" event={"ID":"dfef3acd-6b92-43ee-a0fe-615a0d62ee14","Type":"ContainerStarted","Data":"1ca229a1cd369929c5f65eb193a732b13c1eaf03fd415f7bfd642c16ff3cb70c"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.250653 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.251239 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.751215404 +0000 UTC m=+145.592822859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.251304 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.252076 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.752069094 +0000 UTC m=+145.593676549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.252874 4837 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-nqcvx container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.252953 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" podUID="5858f224-e8b1-4266-b0ab-1d5e72c884c9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.259245 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m78s5" event={"ID":"363b7780-466e-4cff-bee5-8ee51d41d931","Type":"ContainerStarted","Data":"937debc652b51a002b16edb8558b4a2322bb28b151bbc3182e9dd40576ef0dda"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.262839 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" event={"ID":"851e1990-309d-43f3-ab4f-6d18d4f61ebd","Type":"ContainerStarted","Data":"c94d13c5ba4af55a219fd36fbe46e67e8cf0ad6725f65f1a13a60bb98244139b"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.262891 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" event={"ID":"851e1990-309d-43f3-ab4f-6d18d4f61ebd","Type":"ContainerStarted","Data":"e8573fcfb9e225c7b675a4b376d05421ac3b27e91cb500c34f6b8ccb478d3aaa"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.305581 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" event={"ID":"5a0d09c8-5b2a-4f9e-8067-2346e03b1f4c","Type":"ContainerStarted","Data":"0ce95d7a61ae2c4378d15adc4ca19eefdcfcccf5992281f5cb5453d22dd0ac25"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.312113 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" event={"ID":"0ff14da2-35ae-43a1-833f-f79ffbce7f57","Type":"ContainerStarted","Data":"9213d7b1d99bb8c57415a6d095d9ac398b3784374917ca0f50f97bb290111446"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.330105 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" event={"ID":"1ba7ee8e-d52c-403d-af55-25abe87c8c3c","Type":"ContainerStarted","Data":"e40cb9e4128e4099fcea1d79526d9e2517ae74bbf0d42a6d2896cfb9e8c739ff"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.332066 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-25t4p" podStartSLOduration=124.332046971 podStartE2EDuration="2m4.332046971s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.330416952 +0000 UTC m=+145.172024407" watchObservedRunningTime="2025-10-01 07:08:08.332046971 +0000 UTC m=+145.173654426" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.353432 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.354902 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.854879626 +0000 UTC m=+145.696487071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.359845 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rwwps" podStartSLOduration=123.359822553 podStartE2EDuration="2m3.359822553s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.358748807 +0000 UTC m=+145.200356262" watchObservedRunningTime="2025-10-01 07:08:08.359822553 +0000 UTC m=+145.201430008" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.362644 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" event={"ID":"f4ae7b9d-d7f0-4367-860f-81106f4b9d63","Type":"ContainerStarted","Data":"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.382076 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.385493 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" event={"ID":"15fc768b-5e64-4f99-be37-4d7f33158487","Type":"ContainerStarted","Data":"010846fbb5d7c65af7cbb78beb49c0fa54c06e606738d3301b07aa9945030b42"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.388876 4837 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j2drz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.388939 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.420474 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" podStartSLOduration=123.420444248 podStartE2EDuration="2m3.420444248s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.416440824 +0000 UTC m=+145.258048279" watchObservedRunningTime="2025-10-01 07:08:08.420444248 +0000 UTC m=+145.262051703" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.429578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-45m2c" event={"ID":"1985d236-fa0b-441d-824a-82c8be177844","Type":"ContainerStarted","Data":"5e380ffdd7339a941a04969e6b0755eca599f1a013ab88ef7921661f18c8e3f8"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.453165 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" podStartSLOduration=123.453090538 podStartE2EDuration="2m3.453090538s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.447292649 +0000 UTC m=+145.288900104" watchObservedRunningTime="2025-10-01 07:08:08.453090538 +0000 UTC m=+145.294697993" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.455298 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.456801 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:08.956771945 +0000 UTC m=+145.798379400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.453872 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" event={"ID":"880f40fa-c9a4-4fff-a358-4750e9788839","Type":"ContainerStarted","Data":"bdf9a2cee7c5f2e9bb8e111d60fdf47fe6d51bee2da25b89310f92f59c9c3684"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.462462 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" event={"ID":"b281bf11-93c6-4cc2-9f60-9ffe709d7689","Type":"ContainerStarted","Data":"713e645add7cbec77e04cf80bf77c059a0203269399bd7fd59240abf73d3a89e"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.490431 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tmd4d" podStartSLOduration=123.490415927 podStartE2EDuration="2m3.490415927s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.488292707 +0000 UTC m=+145.329900162" watchObservedRunningTime="2025-10-01 07:08:08.490415927 +0000 UTC m=+145.332023382" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.517128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" event={"ID":"87ac3d3f-dc65-4e67-8a79-76d9fbcdece4","Type":"ContainerStarted","Data":"445024b211dd58dda6b92f2e6a8ccafeb24127bbe3bac4ea74d7a5edc667bfbe"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.529061 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" event={"ID":"919c90bd-0d06-4af1-89b2-b3d3aec78ad4","Type":"ContainerStarted","Data":"44ff9bf43f59fef8677997149939effaf2ebab324aaf5af99f85511522b22e82"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.560843 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.561139 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.061118643 +0000 UTC m=+145.902726098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.561619 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.563021 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.063007568 +0000 UTC m=+145.904615013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.566930 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" event={"ID":"065e771b-41ef-4984-aacb-6dd2e1231815","Type":"ContainerStarted","Data":"0e610f916ceae2bb0a3cf0f541ad54b4c262b4355f3cfc2fbf68a11684059536"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.585124 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6q4c5" podStartSLOduration=124.585097485 podStartE2EDuration="2m4.585097485s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.584101641 +0000 UTC m=+145.425709096" watchObservedRunningTime="2025-10-01 07:08:08.585097485 +0000 UTC m=+145.426704940" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.586804 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-45m2c" podStartSLOduration=6.586797245 podStartE2EDuration="6.586797245s" podCreationTimestamp="2025-10-01 07:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.540893491 +0000 UTC m=+145.382500946" watchObservedRunningTime="2025-10-01 07:08:08.586797245 +0000 UTC m=+145.428404700" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.603649 4837 generic.go:334] "Generic (PLEG): container finished" podID="bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5" containerID="5bfafce497cdd9f53fb490a5781accca82cd1eb503cdcf4d2f5ad4e6bec606a1" exitCode=0 Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.603763 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" event={"ID":"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5","Type":"ContainerDied","Data":"5bfafce497cdd9f53fb490a5781accca82cd1eb503cdcf4d2f5ad4e6bec606a1"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.621229 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" event={"ID":"9b4e607c-15af-4ba5-8501-eb50c38c522d","Type":"ContainerStarted","Data":"1f20d801b871da966630c1ede0c6f303a74df2e271cd8d5a03e3809e3a173b4b"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.634104 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9txxm" podStartSLOduration=123.634088783 podStartE2EDuration="2m3.634088783s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.63226757 +0000 UTC m=+145.473875025" watchObservedRunningTime="2025-10-01 07:08:08.634088783 +0000 UTC m=+145.475696238" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.643584 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" event={"ID":"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f","Type":"ContainerStarted","Data":"21daa5eb312c27d90e324cb7999bf5441fce41cfae9550790c61f50f2ec5d641"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.657783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" event={"ID":"6e0a8f48-6289-4c87-aba6-95ae84b691d3","Type":"ContainerStarted","Data":"ed23d00a3cc80f37602e64d38f332548ffdaa33aed27c8fb4de7f900fe8b4338"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.665534 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.665988 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.165954313 +0000 UTC m=+146.007561768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.671020 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.676477 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.176454913 +0000 UTC m=+146.018062368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.694964 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zsd2v" podStartSLOduration=123.694924324 podStartE2EDuration="2m3.694924324s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.684652038 +0000 UTC m=+145.526259493" watchObservedRunningTime="2025-10-01 07:08:08.694924324 +0000 UTC m=+145.536531779" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.700591 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" event={"ID":"51f77b47-e87f-457c-9f20-b61f5f3ea339","Type":"ContainerStarted","Data":"53407123a98fffbbf137006cf4a1326e084d46d9ce281c3b8f023c4cecdbba9d"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.719013 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" event={"ID":"63149934-868e-4a3c-996b-aa7fcdd4cdac","Type":"ContainerStarted","Data":"addaa379a3a929b0ff2d9e5caf3f9b4a04d9d871d3ed9cdfb9f22aae3256d766"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.721015 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.734195 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8mhbv" podStartSLOduration=123.73416476 podStartE2EDuration="2m3.73416476s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.732361306 +0000 UTC m=+145.573968761" watchObservedRunningTime="2025-10-01 07:08:08.73416476 +0000 UTC m=+145.575772215" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.752165 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" event={"ID":"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117","Type":"ContainerStarted","Data":"4eaa95cefbf0e93e9c4e1eeb236af703fb15e897d88aa2c8168363eb2e94af09"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.770857 4837 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2bc7z container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.771344 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" podUID="63149934-868e-4a3c-996b-aa7fcdd4cdac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.771967 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.773468 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.273431356 +0000 UTC m=+146.115038811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.790920 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" event={"ID":"7b8012dc-8975-4a9d-86c3-436e7fca87fa","Type":"ContainerStarted","Data":"dea37e494685172282c83c2fb1d0a43c9964fa28b85cb55a1dbfca6c128f9e4d"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.797799 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" podStartSLOduration=123.797780216 podStartE2EDuration="2m3.797780216s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.795235575 +0000 UTC m=+145.636843030" watchObservedRunningTime="2025-10-01 07:08:08.797780216 +0000 UTC m=+145.639387671" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.814775 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" event={"ID":"6cdca78e-93e0-4e36-94a6-51a65c42ef75","Type":"ContainerStarted","Data":"4d879633d92ed1fa17891525b7eba52d780b061b605d8f75a078eef48d1437fe"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.875301 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.877545 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.377528628 +0000 UTC m=+146.219136073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.904303 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" event={"ID":"1b79aedb-c9cb-46b7-8e26-6b97b8d017af","Type":"ContainerStarted","Data":"facca42cdc7a67a7c04ce680b22fbc2e120b30d698da2d4c5f928861b056fb55"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.906669 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.911590 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" podStartSLOduration=123.911570119 podStartE2EDuration="2m3.911570119s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.903521548 +0000 UTC m=+145.745129013" watchObservedRunningTime="2025-10-01 07:08:08.911570119 +0000 UTC m=+145.753177574" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.932160 4837 generic.go:334] "Generic (PLEG): container finished" podID="4f6c496c-0b1e-4e04-a70c-56ae687b3503" containerID="6a0123b1b860244534527d7683898baa4b4529accb5315db254ae565eb92c638" exitCode=0 Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.933675 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" event={"ID":"4f6c496c-0b1e-4e04-a70c-56ae687b3503","Type":"ContainerDied","Data":"6a0123b1b860244534527d7683898baa4b4529accb5315db254ae565eb92c638"} Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.942863 4837 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mmx44 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.942914 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" podUID="1b79aedb-c9cb-46b7-8e26-6b97b8d017af" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.968311 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" podStartSLOduration=123.968282902 podStartE2EDuration="2m3.968282902s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:08.964595284 +0000 UTC m=+145.806202739" watchObservedRunningTime="2025-10-01 07:08:08.968282902 +0000 UTC m=+145.809890357" Oct 01 07:08:08 crc kubenswrapper[4837]: I1001 07:08:08.976196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:08 crc kubenswrapper[4837]: E1001 07:08:08.981159 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.481138068 +0000 UTC m=+146.322745523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.000030 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.013470 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:09 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:09 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:09 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.013510 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.078310 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.078708 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.578674274 +0000 UTC m=+146.420281729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.149400 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-xszsl" podStartSLOduration=124.14938373 podStartE2EDuration="2m4.14938373s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:09.04912494 +0000 UTC m=+145.890732395" watchObservedRunningTime="2025-10-01 07:08:09.14938373 +0000 UTC m=+145.990991185" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.181683 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.182117 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.68210012 +0000 UTC m=+146.523707575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.246936 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" podStartSLOduration=124.246919156 podStartE2EDuration="2m4.246919156s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:09.24503985 +0000 UTC m=+146.086647305" watchObservedRunningTime="2025-10-01 07:08:09.246919156 +0000 UTC m=+146.088526621" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.287453 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.287808 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.78779669 +0000 UTC m=+146.629404135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.338773 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7qrjz" podStartSLOduration=124.338757035 podStartE2EDuration="2m4.338757035s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:09.32384291 +0000 UTC m=+146.165450365" watchObservedRunningTime="2025-10-01 07:08:09.338757035 +0000 UTC m=+146.180364490" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.388826 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.389467 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.889452495 +0000 UTC m=+146.731059940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.493582 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.493981 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:09.993965026 +0000 UTC m=+146.835572481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.596393 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.596778 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.096762468 +0000 UTC m=+146.938369923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.639967 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cn9dw" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.699424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.700490 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.20047081 +0000 UTC m=+147.042078265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.802139 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.802625 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.302584295 +0000 UTC m=+147.144191760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.904375 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:09 crc kubenswrapper[4837]: E1001 07:08:09.904808 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.404791772 +0000 UTC m=+147.246399227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.944613 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vht5" event={"ID":"7b8012dc-8975-4a9d-86c3-436e7fca87fa","Type":"ContainerStarted","Data":"7cb819e17db832ac711308fb57250665087a3c781f3714bf53ec83459bb08f1a"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.950050 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" event={"ID":"47c7934b-3c05-4fb1-aef3-36f6f2b76d1f","Type":"ContainerStarted","Data":"fe9549bca1593138238af3442a283f1e26604b87ba877ca3eb7891699f8e9704"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.953244 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" event={"ID":"b281bf11-93c6-4cc2-9f60-9ffe709d7689","Type":"ContainerStarted","Data":"627c431ad3722b839304c1dd7bf038d7cf701a5205af442dccb8df979b5d1f34"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.953276 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" event={"ID":"b281bf11-93c6-4cc2-9f60-9ffe709d7689","Type":"ContainerStarted","Data":"1a51d4e3f3a94e525b5895d9b4ae3a186846f37c81a1230e68fe14352de953fc"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.955158 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" event={"ID":"51f77b47-e87f-457c-9f20-b61f5f3ea339","Type":"ContainerStarted","Data":"4fe55f525bc6617b5ab28f8cd687ca5ef17bf84c33dc3c1e82cdc1e8e75467e6"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.955182 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" event={"ID":"51f77b47-e87f-457c-9f20-b61f5f3ea339","Type":"ContainerStarted","Data":"88f787755e7382b4fa8a540952903e25c48302d6a5f691cc046630c0a6397d44"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.956867 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m78s5" event={"ID":"363b7780-466e-4cff-bee5-8ee51d41d931","Type":"ContainerStarted","Data":"b6190a122edd6ba55549a2bc618cbd314efc7b2132bf61b359579b8c8f2af46d"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.956893 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m78s5" event={"ID":"363b7780-466e-4cff-bee5-8ee51d41d931","Type":"ContainerStarted","Data":"ddd3e0cb44daf38c07dbe7ea0254ed98483c262c0e104046f332ef1c1900e928"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.957260 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.958975 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" event={"ID":"15fc768b-5e64-4f99-be37-4d7f33158487","Type":"ContainerStarted","Data":"cc7f5e1b8d3179f85c4572d8f32894adee825f99299386a3ae14f0228adf42b4"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.959354 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.960674 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" event={"ID":"0ff14da2-35ae-43a1-833f-f79ffbce7f57","Type":"ContainerStarted","Data":"713a5f6ce8d0bb4a1a37618a2a0210bc80fbc6d14360f33f3df22fc435a46b6c"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.962107 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hjphq" event={"ID":"6ab1bcc5-281a-4315-81cd-4a803eccfe6f","Type":"ContainerStarted","Data":"290a1900fcb03ab11c3fa625b94d0b3f8a6b92649152e8b54a73ec5cb59ca260"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.964083 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" event={"ID":"9b4e607c-15af-4ba5-8501-eb50c38c522d","Type":"ContainerStarted","Data":"b6d383505e61a3837d62488a36c925e421775c7638457d24caedfff2b66b3ee7"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.966506 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" event={"ID":"785b6d63-389e-4eb8-ba3b-67f23d5dd3c3","Type":"ContainerStarted","Data":"7f9c79672b23bf0c5aa38ab67aacc224086dc9df81bd774a3ae663723d2a227b"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.968933 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" event={"ID":"4f6c496c-0b1e-4e04-a70c-56ae687b3503","Type":"ContainerStarted","Data":"cb6b68e99271ad670adb4815136209f7554c8d576507fa9cdddc0268f3ad8a5b"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.979440 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" event={"ID":"bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5","Type":"ContainerStarted","Data":"43ee0c9ce56ef6ba30dc97db475f5111fd863b626f74b772722d9c961c53ab92"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.979859 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.988264 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" event={"ID":"1ba7ee8e-d52c-403d-af55-25abe87c8c3c","Type":"ContainerStarted","Data":"6ea3f9b92d1f2d1e8c8a94da88fab3394a9633739257a6dd2c1174ce158b5e02"} Oct 01 07:08:09 crc kubenswrapper[4837]: I1001 07:08:09.988307 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" event={"ID":"1ba7ee8e-d52c-403d-af55-25abe87c8c3c","Type":"ContainerStarted","Data":"0e337650b7979ab5710b34a336150bfeb2a33fccfa596db4fbc3a92bd5058264"} Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.002069 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" event={"ID":"dfef3acd-6b92-43ee-a0fe-615a0d62ee14","Type":"ContainerStarted","Data":"ced91a2a78dce58679718108c8fb6aed3d60184bf5b114ce8066d9779aea6e39"} Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.004942 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.006680 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.506663411 +0000 UTC m=+147.348270866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.025706 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:10 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:10 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:10 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.025771 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.044161 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" event={"ID":"63149934-868e-4a3c-996b-aa7fcdd4cdac","Type":"ContainerStarted","Data":"c89caf5a39fca14c1cfbb8fb5e1f40a7e905990c480ccfde2d7d6413cf2599f4"} Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.068876 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" event={"ID":"6e0a8f48-6289-4c87-aba6-95ae84b691d3","Type":"ContainerStarted","Data":"e52683320fab6d9b18c97fa272416a49c5aa79603be6a1cb0b41ecb428e30df9"} Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.068917 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" event={"ID":"6e0a8f48-6289-4c87-aba6-95ae84b691d3","Type":"ContainerStarted","Data":"8cce768b43fce3f38e7007a718e6b3c4ef12dbc5ce1c4462aa3773731f2d69bd"} Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.075800 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.075862 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.078178 4837 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j2drz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.078202 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.109253 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mmx44" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.109516 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nqcvx" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.110448 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ggxrb" podStartSLOduration=125.110435536 podStartE2EDuration="2m5.110435536s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.110363903 +0000 UTC m=+146.951971358" watchObservedRunningTime="2025-10-01 07:08:10.110435536 +0000 UTC m=+146.952042991" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.111175 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z89gn" podStartSLOduration=125.111167013 podStartE2EDuration="2m5.111167013s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.057637627 +0000 UTC m=+146.899245082" watchObservedRunningTime="2025-10-01 07:08:10.111167013 +0000 UTC m=+146.952774468" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.111840 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.129382 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.629364727 +0000 UTC m=+147.470972182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.212765 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.213190 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.713171445 +0000 UTC m=+147.554778890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.243736 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" podStartSLOduration=125.243716304 podStartE2EDuration="2m5.243716304s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.243308073 +0000 UTC m=+147.084915528" watchObservedRunningTime="2025-10-01 07:08:10.243716304 +0000 UTC m=+147.085323759" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.245319 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-m78s5" podStartSLOduration=8.245313802 podStartE2EDuration="8.245313802s" podCreationTimestamp="2025-10-01 07:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.202722356 +0000 UTC m=+147.044329831" watchObservedRunningTime="2025-10-01 07:08:10.245313802 +0000 UTC m=+147.086921257" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.315432 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.315887 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.815869584 +0000 UTC m=+147.657477039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.356610 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rlk6w" podStartSLOduration=125.356589315 podStartE2EDuration="2m5.356589315s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.310124467 +0000 UTC m=+147.151731912" watchObservedRunningTime="2025-10-01 07:08:10.356589315 +0000 UTC m=+147.198196770" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.357346 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" podStartSLOduration=126.357337153 podStartE2EDuration="2m6.357337153s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.353960932 +0000 UTC m=+147.195568377" watchObservedRunningTime="2025-10-01 07:08:10.357337153 +0000 UTC m=+147.198944608" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.403478 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n5kl8" podStartSLOduration=125.403461913 podStartE2EDuration="2m5.403461913s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.400082082 +0000 UTC m=+147.241689537" watchObservedRunningTime="2025-10-01 07:08:10.403461913 +0000 UTC m=+147.245069368" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.418186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.418568 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:10.918552703 +0000 UTC m=+147.760160158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.427551 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" podStartSLOduration=126.427536207 podStartE2EDuration="2m6.427536207s" podCreationTimestamp="2025-10-01 07:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.42597751 +0000 UTC m=+147.267584965" watchObservedRunningTime="2025-10-01 07:08:10.427536207 +0000 UTC m=+147.269143662" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.462720 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4fsxp" podStartSLOduration=125.462702405 podStartE2EDuration="2m5.462702405s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.460944963 +0000 UTC m=+147.302552418" watchObservedRunningTime="2025-10-01 07:08:10.462702405 +0000 UTC m=+147.304309860" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.518163 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hjphq" podStartSLOduration=8.518145767 podStartE2EDuration="8.518145767s" podCreationTimestamp="2025-10-01 07:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.516766494 +0000 UTC m=+147.358373959" watchObservedRunningTime="2025-10-01 07:08:10.518145767 +0000 UTC m=+147.359753222" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.521012 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.521348 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.021334053 +0000 UTC m=+147.862941518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.596621 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" podStartSLOduration=125.596601698 podStartE2EDuration="2m5.596601698s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.594892857 +0000 UTC m=+147.436500312" watchObservedRunningTime="2025-10-01 07:08:10.596601698 +0000 UTC m=+147.438209163" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.598179 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-smwgx" podStartSLOduration=125.598171025 podStartE2EDuration="2m5.598171025s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.548558602 +0000 UTC m=+147.390166057" watchObservedRunningTime="2025-10-01 07:08:10.598171025 +0000 UTC m=+147.439778480" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.622628 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.622787 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.122771021 +0000 UTC m=+147.964378476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.623119 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.623404 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.123397347 +0000 UTC m=+147.965004802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.649595 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z5w2r" podStartSLOduration=125.649571661 podStartE2EDuration="2m5.649571661s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:10.643954197 +0000 UTC m=+147.485561662" watchObservedRunningTime="2025-10-01 07:08:10.649571661 +0000 UTC m=+147.491179116" Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.724181 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.724511 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.224495797 +0000 UTC m=+148.066103252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.826197 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.826759 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.326667333 +0000 UTC m=+148.168274778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.928036 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.928272 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.428239786 +0000 UTC m=+148.269847241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:10 crc kubenswrapper[4837]: I1001 07:08:10.928391 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:10 crc kubenswrapper[4837]: E1001 07:08:10.928822 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.428803619 +0000 UTC m=+148.270411074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.007876 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:11 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:11 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:11 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.008448 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.030293 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.030542 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.530505884 +0000 UTC m=+148.372113349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.030642 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.030711 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.030746 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.048416 4837 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2bc7z container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.048512 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" podUID="63149934-868e-4a3c-996b-aa7fcdd4cdac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.052200 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.055638 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.555614643 +0000 UTC m=+148.397222098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.066277 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.083107 4837 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-6mbk6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.083158 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" podUID="bfa8ee98-4fa7-4fb1-aca5-b72954f27ba5" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.085306 4837 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j2drz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.085383 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.117004 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.117082 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.136731 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.136877 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.63685421 +0000 UTC m=+148.478461675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.137281 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.137801 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.137933 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.142259 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.642250208 +0000 UTC m=+148.483857663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.150252 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.152598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.243973 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.244186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.244554 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.744535917 +0000 UTC m=+148.586143372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.244580 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.253785 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.283919 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2bc7z" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.349736 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.350114 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.850098405 +0000 UTC m=+148.691705860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.399913 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.400863 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.415046 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.443679 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.451212 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.451444 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.451497 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzdpr\" (UniqueName: \"kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.451541 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.451653 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:11.951638125 +0000 UTC m=+148.793245580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.509591 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.510613 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.523746 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555571 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555634 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thcs2\" (UniqueName: \"kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555676 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555748 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555797 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzdpr\" (UniqueName: \"kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555823 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.555851 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.556130 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.056117627 +0000 UTC m=+148.897725082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.556631 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.557188 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.603719 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzdpr\" (UniqueName: \"kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr\") pod \"certified-operators-jtdts\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.649414 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.656986 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.657206 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.657255 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thcs2\" (UniqueName: \"kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.657292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.658294 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.658946 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.158931888 +0000 UTC m=+149.000539343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.661890 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.717728 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.752214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thcs2\" (UniqueName: \"kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2\") pod \"community-operators-njs2f\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.758433 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.758882 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.258866281 +0000 UTC m=+149.100473736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.779235 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.780362 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.841202 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.853880 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.864267 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.864608 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj62r\" (UniqueName: \"kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.864656 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.864742 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.864897 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.364878559 +0000 UTC m=+149.206486014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.901000 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.901964 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.917776 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.967680 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968223 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqdjq\" (UniqueName: \"kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968423 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968444 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968477 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.968512 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj62r\" (UniqueName: \"kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: E1001 07:08:11.969170 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.469156015 +0000 UTC m=+149.310763470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.969796 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:11 crc kubenswrapper[4837]: I1001 07:08:11.970086 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.015389 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj62r\" (UniqueName: \"kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r\") pod \"certified-operators-dcg8q\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.048468 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:12 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:12 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:12 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.048535 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.089618 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.090041 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.090106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.090147 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqdjq\" (UniqueName: \"kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.091330 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.591309608 +0000 UTC m=+149.432917063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.091930 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.092681 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.188776 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqdjq\" (UniqueName: \"kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq\") pod \"community-operators-9k6b2\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.191333 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" event={"ID":"dfef3acd-6b92-43ee-a0fe-615a0d62ee14","Type":"ContainerStarted","Data":"3714b7dfe0705c102fa3db196570c7b7a7dabed580801eaf1edfd6859f09339c"} Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.193329 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.202255 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.702233362 +0000 UTC m=+149.543840817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.305606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.306994 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.80697844 +0000 UTC m=+149.648585895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.328019 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.409814 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.410576 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:12.91055627 +0000 UTC m=+149.752163725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.420325 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.512376 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.513164 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.013139467 +0000 UTC m=+149.854746922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: W1001 07:08:12.566878 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-12abcfba38e1949b9e8c530314c980ee253fe396647bf4e7283112e654b35e16 WatchSource:0}: Error finding container 12abcfba38e1949b9e8c530314c980ee253fe396647bf4e7283112e654b35e16: Status 404 returned error can't find the container with id 12abcfba38e1949b9e8c530314c980ee253fe396647bf4e7283112e654b35e16 Oct 01 07:08:12 crc kubenswrapper[4837]: W1001 07:08:12.592996 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-0ba7f340ea520074c1329a09ccbf256e505d333ecbc2e8493c47f420d899f720 WatchSource:0}: Error finding container 0ba7f340ea520074c1329a09ccbf256e505d333ecbc2e8493c47f420d899f720: Status 404 returned error can't find the container with id 0ba7f340ea520074c1329a09ccbf256e505d333ecbc2e8493c47f420d899f720 Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.616608 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.617339 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.117327521 +0000 UTC m=+149.958934976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.729229 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.729619 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.229604448 +0000 UTC m=+150.071211893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.832498 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.832857 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.332841539 +0000 UTC m=+150.174448994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.934516 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.940400 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.440372194 +0000 UTC m=+150.281979649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.940949 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:12 crc kubenswrapper[4837]: E1001 07:08:12.941406 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.441387627 +0000 UTC m=+150.282995072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:12 crc kubenswrapper[4837]: I1001 07:08:12.977355 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.019191 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:13 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:13 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:13 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.019258 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.042672 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.042889 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.542855027 +0000 UTC m=+150.384462482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.043182 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.043632 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.543618865 +0000 UTC m=+150.385226320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.067858 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.070734 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.071526 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.074212 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.076410 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.076790 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.144064 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.144486 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.144547 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.144653 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.644637324 +0000 UTC m=+150.486244779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.198597 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"12abcfba38e1949b9e8c530314c980ee253fe396647bf4e7283112e654b35e16"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.208194 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7a7ebe6fde3a7a00bb559967371362fc5623231019b7cecbdaf74a6f28f1f588"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.209888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerStarted","Data":"2b9d2b45ebbd7c2c24a89596188edb68a51aaeb6679fb1ff4eff721b1bfb5d62"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.211004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerStarted","Data":"aa8ee3f8923adac7cd5c99742d506025d47eafaaf066c6bd82c9ccde067028c8"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.215474 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" event={"ID":"dfef3acd-6b92-43ee-a0fe-615a0d62ee14","Type":"ContainerStarted","Data":"591b3479a4fe797661b1c3d2dbd9e2f2882c7aeb511f51499a078049cfecdf6e"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.229942 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0ba7f340ea520074c1329a09ccbf256e505d333ecbc2e8493c47f420d899f720"} Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.247726 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.247824 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.247868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.248233 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.748218824 +0000 UTC m=+150.589826279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.249115 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.272006 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.278912 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.279908 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.283546 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.296401 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.349112 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.349302 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.349467 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.849407667 +0000 UTC m=+150.691015122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.349800 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.349895 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.349980 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t76zg\" (UniqueName: \"kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.350290 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.850274997 +0000 UTC m=+150.691882452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.387779 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:08:13 crc kubenswrapper[4837]: W1001 07:08:13.395525 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b25ef2f_3713_4377_8e5c_c0e9ef894969.slice/crio-13b1ec5f930cc47077d6b5dc5f9d51229591e083aa5bc899378e232af149bee7 WatchSource:0}: Error finding container 13b1ec5f930cc47077d6b5dc5f9d51229591e083aa5bc899378e232af149bee7: Status 404 returned error can't find the container with id 13b1ec5f930cc47077d6b5dc5f9d51229591e083aa5bc899378e232af149bee7 Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.396737 4837 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.451417 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.451713 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.451760 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.451803 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t76zg\" (UniqueName: \"kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.451916 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:13.95188533 +0000 UTC m=+150.793492805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.452463 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.454477 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.477771 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t76zg\" (UniqueName: \"kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg\") pod \"redhat-marketplace-qhvxs\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.521929 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.549786 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.552659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.553039 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.053025721 +0000 UTC m=+150.894633176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.605594 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.653867 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.654117 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.154078901 +0000 UTC m=+150.995686366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.655614 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.658224 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.158196729 +0000 UTC m=+150.999804364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.679051 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.683451 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.690470 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:08:13 crc kubenswrapper[4837]: W1001 07:08:13.695867 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d2ea16a_c2ec_4f68_9c32_812598caac93.slice/crio-c819a38c1981748d6da6d2ea9ac403bc172988002096da15e812d93e95056090 WatchSource:0}: Error finding container c819a38c1981748d6da6d2ea9ac403bc172988002096da15e812d93e95056090: Status 404 returned error can't find the container with id c819a38c1981748d6da6d2ea9ac403bc172988002096da15e812d93e95056090 Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.756779 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.757208 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.256998355 +0000 UTC m=+151.098605820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.758255 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97b6t\" (UniqueName: \"kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.758323 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.758363 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.758570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.758765 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.258751707 +0000 UTC m=+151.100359162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.768820 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 07:08:13 crc kubenswrapper[4837]: W1001 07:08:13.792072 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod548ce95e_d2b9_44f9_bc8f_c0cda36710f7.slice/crio-480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a WatchSource:0}: Error finding container 480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a: Status 404 returned error can't find the container with id 480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.853373 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.860620 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.861223 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.361199349 +0000 UTC m=+151.202806804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.861324 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.861390 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97b6t\" (UniqueName: \"kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.861438 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.861490 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.862018 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.862047 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.3620378 +0000 UTC m=+151.203645255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.862193 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.893904 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97b6t\" (UniqueName: \"kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t\") pod \"redhat-marketplace-rbl6p\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.895270 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6mbk6" Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.963050 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.963277 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.463195592 +0000 UTC m=+151.304803047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.963526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:13 crc kubenswrapper[4837]: E1001 07:08:13.965137 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 07:08:14.465119488 +0000 UTC m=+151.306726943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vxkzq" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 07:08:13 crc kubenswrapper[4837]: I1001 07:08:13.997601 4837 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-01T07:08:13.396764806Z","Handler":null,"Name":""} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.004781 4837 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.004821 4837 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.007081 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.017273 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:14 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:14 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:14 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.018201 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.064434 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.089721 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.165793 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.173481 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.173543 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.199770 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vxkzq\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.264741 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" event={"ID":"dfef3acd-6b92-43ee-a0fe-615a0d62ee14","Type":"ContainerStarted","Data":"e0a05e8a3598087cd533b46c34942fd5d36658bbb51d67d2964bc237e7df405b"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.271864 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"548ce95e-d2b9-44f9-bc8f-c0cda36710f7","Type":"ContainerStarted","Data":"8eee67b126dfdeab2c3665ba4b34a06212a3d52713c92e4464397c1f1a2b2ef7"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.271925 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"548ce95e-d2b9-44f9-bc8f-c0cda36710f7","Type":"ContainerStarted","Data":"480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.277343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"108660b125de2b772f771bffb80b6a4279fd47fd97c3dd3d55a96e8b9f6df419"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.277602 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.279596 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" containerID="4eaa95cefbf0e93e9c4e1eeb236af703fb15e897d88aa2c8168363eb2e94af09" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.279684 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" event={"ID":"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117","Type":"ContainerDied","Data":"4eaa95cefbf0e93e9c4e1eeb236af703fb15e897d88aa2c8168363eb2e94af09"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.281290 4837 generic.go:334] "Generic (PLEG): container finished" podID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerID="3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.281407 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerDied","Data":"3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.281481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerStarted","Data":"4090276c9f2b12a01fdc1dc13b42f3da82f2064ab156ceab9fb953a4b934ef1d"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.286762 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.292025 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerID="a46d0c951ce2ec470d7480e23b9380e1bbd4e7fae5ffda14e5d463d81c2c39d6" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.292135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerDied","Data":"a46d0c951ce2ec470d7480e23b9380e1bbd4e7fae5ffda14e5d463d81c2c39d6"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.292175 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerStarted","Data":"13b1ec5f930cc47077d6b5dc5f9d51229591e083aa5bc899378e232af149bee7"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.302052 4837 generic.go:334] "Generic (PLEG): container finished" podID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerID="dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.302128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerDied","Data":"dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.302156 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerStarted","Data":"c819a38c1981748d6da6d2ea9ac403bc172988002096da15e812d93e95056090"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.313128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7badce1e4533b2bdd16371b2f32d41709b00e28c0e63bf98ee2f4dd60e66054e"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.322458 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9lclf" podStartSLOduration=12.322432298 podStartE2EDuration="12.322432298s" podCreationTimestamp="2025-10-01 07:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:14.307393559 +0000 UTC m=+151.149001004" watchObservedRunningTime="2025-10-01 07:08:14.322432298 +0000 UTC m=+151.164039753" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.334856 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"222b14d46d96e45b85d16833bf22d3be46e9f3fe3551b338102ac973db41f513"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.337683 4837 generic.go:334] "Generic (PLEG): container finished" podID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerID="ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.337772 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerDied","Data":"ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.339608 4837 generic.go:334] "Generic (PLEG): container finished" podID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerID="e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5" exitCode=0 Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.339660 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerDied","Data":"e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5"} Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.398794 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.398770847 podStartE2EDuration="1.398770847s" podCreationTimestamp="2025-10-01 07:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:14.396534735 +0000 UTC m=+151.238142190" watchObservedRunningTime="2025-10-01 07:08:14.398770847 +0000 UTC m=+151.240378312" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.500201 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.534232 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:08:14 crc kubenswrapper[4837]: W1001 07:08:14.559886 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94ab3cd8_ecf2_454d_99e9_31ccc57f77ef.slice/crio-593b843a3651cbf81bdbb133ddc907f8cba67c61a5a831b8096ecd9868c6d40c WatchSource:0}: Error finding container 593b843a3651cbf81bdbb133ddc907f8cba67c61a5a831b8096ecd9868c6d40c: Status 404 returned error can't find the container with id 593b843a3651cbf81bdbb133ddc907f8cba67c61a5a831b8096ecd9868c6d40c Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.629331 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.629730 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.645035 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.688748 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.690021 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.694146 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.698003 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.728752 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.730380 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.745521 4837 patch_prober.go:28] interesting pod/console-f9d7485db-mg4h6 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.745608 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mg4h6" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.776629 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.776765 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.776841 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhvmr\" (UniqueName: \"kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.781679 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.829269 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.829327 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.831308 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.831403 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:14 crc kubenswrapper[4837]: E1001 07:08:14.867135 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod548ce95e_d2b9_44f9_bc8f_c0cda36710f7.slice/crio-8eee67b126dfdeab2c3665ba4b34a06212a3d52713c92e4464397c1f1a2b2ef7.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.877802 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.878210 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.878212 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.878234 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhvmr\" (UniqueName: \"kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.878437 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.899060 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhvmr\" (UniqueName: \"kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr\") pod \"redhat-operators-t5ql6\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.899297 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.899770 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:14 crc kubenswrapper[4837]: I1001 07:08:14.906307 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.003275 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.008150 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:15 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:15 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:15 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.008562 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.092528 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.094076 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.097824 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.110984 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.182986 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.183071 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.183242 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dqrz\" (UniqueName: \"kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.286476 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dqrz\" (UniqueName: \"kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.287010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.287083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.287519 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.287553 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.322060 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dqrz\" (UniqueName: \"kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz\") pod \"redhat-operators-q8cdq\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.356817 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" event={"ID":"56966f17-7502-47e9-a482-0e9f3fc63b0d","Type":"ContainerStarted","Data":"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335"} Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.356882 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" event={"ID":"56966f17-7502-47e9-a482-0e9f3fc63b0d","Type":"ContainerStarted","Data":"3a1023e97dc05b25448bf7ca555f2bde5e2f6a05d89c6218b43fd3347941aad0"} Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.357153 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.366925 4837 generic.go:334] "Generic (PLEG): container finished" podID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerID="4202eb23691d964facc3aecb10fe197a2e9571cb679a2333b0a294f7f752e468" exitCode=0 Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.367654 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerDied","Data":"4202eb23691d964facc3aecb10fe197a2e9571cb679a2333b0a294f7f752e468"} Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.367682 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerStarted","Data":"593b843a3651cbf81bdbb133ddc907f8cba67c61a5a831b8096ecd9868c6d40c"} Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.377514 4837 generic.go:334] "Generic (PLEG): container finished" podID="548ce95e-d2b9-44f9-bc8f-c0cda36710f7" containerID="8eee67b126dfdeab2c3665ba4b34a06212a3d52713c92e4464397c1f1a2b2ef7" exitCode=0 Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.378481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"548ce95e-d2b9-44f9-bc8f-c0cda36710f7","Type":"ContainerDied","Data":"8eee67b126dfdeab2c3665ba4b34a06212a3d52713c92e4464397c1f1a2b2ef7"} Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.387123 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" podStartSLOduration=130.387098664 podStartE2EDuration="2m10.387098664s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:15.386569941 +0000 UTC m=+152.228177396" watchObservedRunningTime="2025-10-01 07:08:15.387098664 +0000 UTC m=+152.228706119" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.388942 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bftpg" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.402309 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rklvs" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.415796 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.448938 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.455570 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.456571 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.459210 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.478481 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.481492 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 07:08:15 crc kubenswrapper[4837]: W1001 07:08:15.494961 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod533b6b0d_129b_4d94_b233_541162a9a58f.slice/crio-f82cee67614c4fd010b0ae1ff6320e5fa8776cd729a25b5ca1ca61c57f5a6c3e WatchSource:0}: Error finding container f82cee67614c4fd010b0ae1ff6320e5fa8776cd729a25b5ca1ca61c57f5a6c3e: Status 404 returned error can't find the container with id f82cee67614c4fd010b0ae1ff6320e5fa8776cd729a25b5ca1ca61c57f5a6c3e Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.498344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.498956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.609344 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.610233 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.610391 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.663475 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.674727 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.766504 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.801744 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.838977 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.918427 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nswm\" (UniqueName: \"kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm\") pod \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.918512 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume\") pod \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.918593 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume\") pod \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\" (UID: \"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117\") " Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.924231 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume" (OuterVolumeSpecName: "config-volume") pod "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" (UID: "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.940945 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" (UID: "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:08:15 crc kubenswrapper[4837]: I1001 07:08:15.946147 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm" (OuterVolumeSpecName: "kube-api-access-5nswm") pod "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" (UID: "b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117"). InnerVolumeSpecName "kube-api-access-5nswm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.010892 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:16 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:16 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:16 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.010948 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.021424 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.021466 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.021478 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nswm\" (UniqueName: \"kubernetes.io/projected/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117-kube-api-access-5nswm\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.087719 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.330170 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.390285 4837 generic.go:334] "Generic (PLEG): container finished" podID="533b6b0d-129b-4d94-b233-541162a9a58f" containerID="2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f" exitCode=0 Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.390355 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerDied","Data":"2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.390385 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerStarted","Data":"f82cee67614c4fd010b0ae1ff6320e5fa8776cd729a25b5ca1ca61c57f5a6c3e"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.408282 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d628bd9d-69b8-48fe-962f-70986bf17703","Type":"ContainerStarted","Data":"5bd820fe8c8fcbbddbbb81fa5f2a641499dce18d93a03cf8c3626426de2f06a3"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.424204 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.426243 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l" event={"ID":"b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117","Type":"ContainerDied","Data":"e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.426300 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6e7609a60b7b475480a8103679c12f386fab997db8b0d5b766f4da73e1afec0" Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.461074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerStarted","Data":"b0e25a27de32a72378f2925c579171430bee97bd557235a90c3cba115093b649"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.461128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerStarted","Data":"099b47bd075040f668f132dfd5c131c9147da3cc5bc9fd3a924c9f687efb8496"} Oct 01 07:08:16 crc kubenswrapper[4837]: I1001 07:08:16.878405 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.006019 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:17 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:17 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:17 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.006084 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.038512 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir\") pod \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.038656 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access\") pod \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\" (UID: \"548ce95e-d2b9-44f9-bc8f-c0cda36710f7\") " Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.040177 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "548ce95e-d2b9-44f9-bc8f-c0cda36710f7" (UID: "548ce95e-d2b9-44f9-bc8f-c0cda36710f7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.054546 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "548ce95e-d2b9-44f9-bc8f-c0cda36710f7" (UID: "548ce95e-d2b9-44f9-bc8f-c0cda36710f7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.141389 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.141421 4837 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/548ce95e-d2b9-44f9-bc8f-c0cda36710f7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.499949 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerID="b0e25a27de32a72378f2925c579171430bee97bd557235a90c3cba115093b649" exitCode=0 Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.500078 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerDied","Data":"b0e25a27de32a72378f2925c579171430bee97bd557235a90c3cba115093b649"} Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.512139 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"548ce95e-d2b9-44f9-bc8f-c0cda36710f7","Type":"ContainerDied","Data":"480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a"} Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.512173 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="480acbb97f5c08c6ef6d06c84499a664023cf8a1c76c29ede8719a38aa605e3a" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.512227 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.577671 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d628bd9d-69b8-48fe-962f-70986bf17703","Type":"ContainerStarted","Data":"92b7a1e25b807f93f7242214d1573976ee39cd94f53af4ce73583cced16b9160"} Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.615031 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.615013356 podStartE2EDuration="2.615013356s" podCreationTimestamp="2025-10-01 07:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:17.611905442 +0000 UTC m=+154.453512897" watchObservedRunningTime="2025-10-01 07:08:17.615013356 +0000 UTC m=+154.456620811" Oct 01 07:08:17 crc kubenswrapper[4837]: I1001 07:08:17.745036 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:08:18 crc kubenswrapper[4837]: I1001 07:08:18.011999 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:18 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:18 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:18 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:18 crc kubenswrapper[4837]: I1001 07:08:18.012070 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:18 crc kubenswrapper[4837]: I1001 07:08:18.615945 4837 generic.go:334] "Generic (PLEG): container finished" podID="d628bd9d-69b8-48fe-962f-70986bf17703" containerID="92b7a1e25b807f93f7242214d1573976ee39cd94f53af4ce73583cced16b9160" exitCode=0 Oct 01 07:08:18 crc kubenswrapper[4837]: I1001 07:08:18.616003 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d628bd9d-69b8-48fe-962f-70986bf17703","Type":"ContainerDied","Data":"92b7a1e25b807f93f7242214d1573976ee39cd94f53af4ce73583cced16b9160"} Oct 01 07:08:19 crc kubenswrapper[4837]: I1001 07:08:19.005270 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:19 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:19 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:19 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:19 crc kubenswrapper[4837]: I1001 07:08:19.005340 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.028129 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:20 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:20 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:20 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.028430 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.159672 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.311896 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir\") pod \"d628bd9d-69b8-48fe-962f-70986bf17703\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.311964 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access\") pod \"d628bd9d-69b8-48fe-962f-70986bf17703\" (UID: \"d628bd9d-69b8-48fe-962f-70986bf17703\") " Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.312065 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d628bd9d-69b8-48fe-962f-70986bf17703" (UID: "d628bd9d-69b8-48fe-962f-70986bf17703"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.312292 4837 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d628bd9d-69b8-48fe-962f-70986bf17703-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.318064 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d628bd9d-69b8-48fe-962f-70986bf17703" (UID: "d628bd9d-69b8-48fe-962f-70986bf17703"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.414146 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d628bd9d-69b8-48fe-962f-70986bf17703-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.634594 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d628bd9d-69b8-48fe-962f-70986bf17703","Type":"ContainerDied","Data":"5bd820fe8c8fcbbddbbb81fa5f2a641499dce18d93a03cf8c3626426de2f06a3"} Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.634659 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.634702 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd820fe8c8fcbbddbbb81fa5f2a641499dce18d93a03cf8c3626426de2f06a3" Oct 01 07:08:20 crc kubenswrapper[4837]: I1001 07:08:20.760610 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-m78s5" Oct 01 07:08:21 crc kubenswrapper[4837]: I1001 07:08:21.011966 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:21 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:21 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:21 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:21 crc kubenswrapper[4837]: I1001 07:08:21.012043 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:22 crc kubenswrapper[4837]: I1001 07:08:22.005898 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:22 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:22 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:22 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:22 crc kubenswrapper[4837]: I1001 07:08:22.005968 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:23 crc kubenswrapper[4837]: I1001 07:08:23.005901 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:23 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:23 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:23 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:23 crc kubenswrapper[4837]: I1001 07:08:23.005960 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:23 crc kubenswrapper[4837]: I1001 07:08:23.079526 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:08:23 crc kubenswrapper[4837]: I1001 07:08:23.079582 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.004981 4837 patch_prober.go:28] interesting pod/router-default-5444994796-vb728 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 07:08:24 crc kubenswrapper[4837]: [-]has-synced failed: reason withheld Oct 01 07:08:24 crc kubenswrapper[4837]: [+]process-running ok Oct 01 07:08:24 crc kubenswrapper[4837]: healthz check failed Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.005033 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vb728" podUID="7ca8ccaf-fa72-4f5d-ba14-721c90583efb" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.711599 4837 patch_prober.go:28] interesting pod/console-f9d7485db-mg4h6 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.712001 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mg4h6" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.829939 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.829996 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.831047 4837 patch_prober.go:28] interesting pod/downloads-7954f5f757-25t4p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 01 07:08:24 crc kubenswrapper[4837]: I1001 07:08:24.831069 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-25t4p" podUID="5be4a157-1e4f-4e93-8f85-2c01e1f60ff1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 01 07:08:25 crc kubenswrapper[4837]: I1001 07:08:25.006649 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:25 crc kubenswrapper[4837]: I1001 07:08:25.009229 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-vb728" Oct 01 07:08:27 crc kubenswrapper[4837]: I1001 07:08:27.048183 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:08:27 crc kubenswrapper[4837]: I1001 07:08:27.066847 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/95803be9-8a27-4569-958d-81666ad8defc-metrics-certs\") pod \"network-metrics-daemon-z66wh\" (UID: \"95803be9-8a27-4569-958d-81666ad8defc\") " pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:08:27 crc kubenswrapper[4837]: I1001 07:08:27.233746 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z66wh" Oct 01 07:08:34 crc kubenswrapper[4837]: I1001 07:08:34.512667 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:08:34 crc kubenswrapper[4837]: I1001 07:08:34.715715 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:34 crc kubenswrapper[4837]: I1001 07:08:34.720292 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:08:34 crc kubenswrapper[4837]: I1001 07:08:34.844650 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-25t4p" Oct 01 07:08:45 crc kubenswrapper[4837]: I1001 07:08:45.623552 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sl6hv" Oct 01 07:08:46 crc kubenswrapper[4837]: E1001 07:08:46.006120 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 07:08:46 crc kubenswrapper[4837]: E1001 07:08:46.006377 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hzdpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jtdts_openshift-marketplace(9a25e900-33fa-4d02-8ea5-377911d8e4e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:46 crc kubenswrapper[4837]: E1001 07:08:46.007907 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jtdts" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" Oct 01 07:08:47 crc kubenswrapper[4837]: E1001 07:08:47.215492 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jtdts" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" Oct 01 07:08:47 crc kubenswrapper[4837]: E1001 07:08:47.960799 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 07:08:47 crc kubenswrapper[4837]: E1001 07:08:47.961038 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t76zg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qhvxs_openshift-marketplace(534df8e2-faff-47cf-a3d4-e62505fab14a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:47 crc kubenswrapper[4837]: E1001 07:08:47.962304 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qhvxs" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" Oct 01 07:08:48 crc kubenswrapper[4837]: E1001 07:08:48.129216 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 07:08:48 crc kubenswrapper[4837]: E1001 07:08:48.129953 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xj62r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dcg8q_openshift-marketplace(3d2ea16a-c2ec-4f68-9c32-812598caac93): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:48 crc kubenswrapper[4837]: E1001 07:08:48.131119 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dcg8q" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" Oct 01 07:08:49 crc kubenswrapper[4837]: E1001 07:08:49.262238 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qhvxs" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" Oct 01 07:08:49 crc kubenswrapper[4837]: E1001 07:08:49.262587 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dcg8q" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" Oct 01 07:08:49 crc kubenswrapper[4837]: E1001 07:08:49.395825 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 07:08:49 crc kubenswrapper[4837]: E1001 07:08:49.396188 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqdjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9k6b2_openshift-marketplace(6b25ef2f-3713-4377-8e5c-c0e9ef894969): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:49 crc kubenswrapper[4837]: E1001 07:08:49.397497 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9k6b2" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" Oct 01 07:08:51 crc kubenswrapper[4837]: I1001 07:08:51.254444 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.378800 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9k6b2" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.463599 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.463837 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-97b6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rbl6p_openshift-marketplace(94ab3cd8-ecf2-454d-99e9-31ccc57f77ef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.466174 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rbl6p" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.476382 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.476569 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hhvmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-t5ql6_openshift-marketplace(533b6b0d-129b-4d94-b233-541162a9a58f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.478063 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-t5ql6" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.510586 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.511111 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thcs2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-njs2f_openshift-marketplace(5adadfe2-e89d-43bf-94a4-b3cdb58efaa2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.513565 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-njs2f" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.541206 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.541356 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4dqrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-q8cdq_openshift-marketplace(7b94d903-fa51-451e-b95a-b251ad3302ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.542655 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-q8cdq" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" Oct 01 07:08:52 crc kubenswrapper[4837]: I1001 07:08:52.803438 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z66wh"] Oct 01 07:08:52 crc kubenswrapper[4837]: I1001 07:08:52.894045 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z66wh" event={"ID":"95803be9-8a27-4569-958d-81666ad8defc","Type":"ContainerStarted","Data":"79d76664c7427c66340780756a28a90819ad62663df9c6afa5d765d170fc9a37"} Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.896020 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rbl6p" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.897286 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-t5ql6" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.897381 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-njs2f" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" Oct 01 07:08:52 crc kubenswrapper[4837]: E1001 07:08:52.896662 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-q8cdq" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" Oct 01 07:08:53 crc kubenswrapper[4837]: I1001 07:08:53.079495 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:08:53 crc kubenswrapper[4837]: I1001 07:08:53.079669 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:08:53 crc kubenswrapper[4837]: I1001 07:08:53.902378 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z66wh" event={"ID":"95803be9-8a27-4569-958d-81666ad8defc","Type":"ContainerStarted","Data":"c92cfb1be6925cbe70ed86e492f7d8ee0afa03efc83cad73291b7a90d475093a"} Oct 01 07:08:53 crc kubenswrapper[4837]: I1001 07:08:53.902768 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z66wh" event={"ID":"95803be9-8a27-4569-958d-81666ad8defc","Type":"ContainerStarted","Data":"eeba1b1afff93d170bbcc397a5c9c37e311733fc74c3bbd9c1a722eb9a299f7f"} Oct 01 07:08:53 crc kubenswrapper[4837]: I1001 07:08:53.926817 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-z66wh" podStartSLOduration=168.926772296 podStartE2EDuration="2m48.926772296s" podCreationTimestamp="2025-10-01 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:08:53.924750807 +0000 UTC m=+190.766358292" watchObservedRunningTime="2025-10-01 07:08:53.926772296 +0000 UTC m=+190.768379761" Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.970439 4837 generic.go:334] "Generic (PLEG): container finished" podID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerID="1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd" exitCode=0 Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.970543 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerDied","Data":"1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd"} Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.975116 4837 generic.go:334] "Generic (PLEG): container finished" podID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerID="21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6" exitCode=0 Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.975294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerDied","Data":"21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6"} Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.988827 4837 generic.go:334] "Generic (PLEG): container finished" podID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerID="d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42" exitCode=0 Oct 01 07:09:02 crc kubenswrapper[4837]: I1001 07:09:02.988937 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerDied","Data":"d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42"} Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.004138 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerStarted","Data":"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c"} Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.011264 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerStarted","Data":"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd"} Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.014569 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerStarted","Data":"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb"} Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.031356 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jtdts" podStartSLOduration=4.403122055 podStartE2EDuration="54.031340276s" podCreationTimestamp="2025-10-01 07:08:11 +0000 UTC" firstStartedPulling="2025-10-01 07:08:14.341512462 +0000 UTC m=+151.183119917" lastFinishedPulling="2025-10-01 07:09:03.969730653 +0000 UTC m=+200.811338138" observedRunningTime="2025-10-01 07:09:05.027208338 +0000 UTC m=+201.868815803" watchObservedRunningTime="2025-10-01 07:09:05.031340276 +0000 UTC m=+201.872947731" Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.054871 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qhvxs" podStartSLOduration=2.356262409 podStartE2EDuration="52.054854937s" podCreationTimestamp="2025-10-01 07:08:13 +0000 UTC" firstStartedPulling="2025-10-01 07:08:14.286247055 +0000 UTC m=+151.127854510" lastFinishedPulling="2025-10-01 07:09:03.984839543 +0000 UTC m=+200.826447038" observedRunningTime="2025-10-01 07:09:05.052569822 +0000 UTC m=+201.894177277" watchObservedRunningTime="2025-10-01 07:09:05.054854937 +0000 UTC m=+201.896462392" Oct 01 07:09:05 crc kubenswrapper[4837]: I1001 07:09:05.072420 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dcg8q" podStartSLOduration=4.43683749 podStartE2EDuration="54.072398995s" podCreationTimestamp="2025-10-01 07:08:11 +0000 UTC" firstStartedPulling="2025-10-01 07:08:14.309865268 +0000 UTC m=+151.151472723" lastFinishedPulling="2025-10-01 07:09:03.945426733 +0000 UTC m=+200.787034228" observedRunningTime="2025-10-01 07:09:05.069217259 +0000 UTC m=+201.910824724" watchObservedRunningTime="2025-10-01 07:09:05.072398995 +0000 UTC m=+201.914006450" Oct 01 07:09:06 crc kubenswrapper[4837]: I1001 07:09:06.022927 4837 generic.go:334] "Generic (PLEG): container finished" podID="533b6b0d-129b-4d94-b233-541162a9a58f" containerID="a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c" exitCode=0 Oct 01 07:09:06 crc kubenswrapper[4837]: I1001 07:09:06.022994 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerDied","Data":"a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c"} Oct 01 07:09:07 crc kubenswrapper[4837]: I1001 07:09:07.031826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerStarted","Data":"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709"} Oct 01 07:09:07 crc kubenswrapper[4837]: I1001 07:09:07.036017 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerStarted","Data":"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569"} Oct 01 07:09:07 crc kubenswrapper[4837]: I1001 07:09:07.042549 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerID="da91dfb0eb1dce808addcdacf26db4161218aebbfe6e08c117e61b0daed8891c" exitCode=0 Oct 01 07:09:07 crc kubenswrapper[4837]: I1001 07:09:07.042629 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerDied","Data":"da91dfb0eb1dce808addcdacf26db4161218aebbfe6e08c117e61b0daed8891c"} Oct 01 07:09:07 crc kubenswrapper[4837]: I1001 07:09:07.111501 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t5ql6" podStartSLOduration=3.023194515 podStartE2EDuration="53.111478755s" podCreationTimestamp="2025-10-01 07:08:14 +0000 UTC" firstStartedPulling="2025-10-01 07:08:16.394100935 +0000 UTC m=+153.235708390" lastFinishedPulling="2025-10-01 07:09:06.482385165 +0000 UTC m=+203.323992630" observedRunningTime="2025-10-01 07:09:07.084222136 +0000 UTC m=+203.925829631" watchObservedRunningTime="2025-10-01 07:09:07.111478755 +0000 UTC m=+203.953086230" Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.055760 4837 generic.go:334] "Generic (PLEG): container finished" podID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerID="973a4d13cefd9888ba95a4abe0c8c5f917926a2ff97637f2b1afdc6daa4ea9f5" exitCode=0 Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.056242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerDied","Data":"973a4d13cefd9888ba95a4abe0c8c5f917926a2ff97637f2b1afdc6daa4ea9f5"} Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.065339 4837 generic.go:334] "Generic (PLEG): container finished" podID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerID="37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709" exitCode=0 Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.065435 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerDied","Data":"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709"} Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.074857 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerStarted","Data":"2055a4807cf966100b6adf22bca618282ee42448ddc59491151dd1e3cd335887"} Oct 01 07:09:08 crc kubenswrapper[4837]: I1001 07:09:08.131871 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9k6b2" podStartSLOduration=3.795716332 podStartE2EDuration="57.131850725s" podCreationTimestamp="2025-10-01 07:08:11 +0000 UTC" firstStartedPulling="2025-10-01 07:08:14.295655079 +0000 UTC m=+151.137262534" lastFinishedPulling="2025-10-01 07:09:07.631789462 +0000 UTC m=+204.473396927" observedRunningTime="2025-10-01 07:09:08.13119764 +0000 UTC m=+204.972805125" watchObservedRunningTime="2025-10-01 07:09:08.131850725 +0000 UTC m=+204.973458180" Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.084419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerStarted","Data":"79184f87785f55c775fab808050fee50981575325abd70106cae97b6a7765196"} Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.089191 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerID="787a48808d58c3dd4d262779086efe9acacfb45bb827fc45512511cbf62c6247" exitCode=0 Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.089242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerDied","Data":"787a48808d58c3dd4d262779086efe9acacfb45bb827fc45512511cbf62c6247"} Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.092856 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerStarted","Data":"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39"} Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.119187 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rbl6p" podStartSLOduration=2.799656332 podStartE2EDuration="56.119156177s" podCreationTimestamp="2025-10-01 07:08:13 +0000 UTC" firstStartedPulling="2025-10-01 07:08:15.372320072 +0000 UTC m=+152.213927527" lastFinishedPulling="2025-10-01 07:09:08.691819917 +0000 UTC m=+205.533427372" observedRunningTime="2025-10-01 07:09:09.116182856 +0000 UTC m=+205.957790321" watchObservedRunningTime="2025-10-01 07:09:09.119156177 +0000 UTC m=+205.960763632" Oct 01 07:09:09 crc kubenswrapper[4837]: I1001 07:09:09.142560 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-njs2f" podStartSLOduration=3.870606857 podStartE2EDuration="58.142531514s" podCreationTimestamp="2025-10-01 07:08:11 +0000 UTC" firstStartedPulling="2025-10-01 07:08:14.339189607 +0000 UTC m=+151.180797072" lastFinishedPulling="2025-10-01 07:09:08.611114274 +0000 UTC m=+205.452721729" observedRunningTime="2025-10-01 07:09:09.140235079 +0000 UTC m=+205.981842544" watchObservedRunningTime="2025-10-01 07:09:09.142531514 +0000 UTC m=+205.984138979" Oct 01 07:09:11 crc kubenswrapper[4837]: I1001 07:09:11.718520 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:09:11 crc kubenswrapper[4837]: I1001 07:09:11.718899 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:09:11 crc kubenswrapper[4837]: I1001 07:09:11.855183 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:09:11 crc kubenswrapper[4837]: I1001 07:09:11.855634 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.329088 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.331684 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.347913 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.348033 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.389217 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.413933 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.421329 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.421908 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:12 crc kubenswrapper[4837]: I1001 07:09:12.464714 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.200021 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.201823 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.201858 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.607075 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.607745 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:09:13 crc kubenswrapper[4837]: I1001 07:09:13.713149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.007473 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.007530 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.072548 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.137981 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerStarted","Data":"114bc921be3381250b9e3e71ea03a62efae015295524a9cda3a56aa4b6e8f58a"} Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.167369 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8cdq" podStartSLOduration=3.555431211 podStartE2EDuration="59.167344558s" podCreationTimestamp="2025-10-01 07:08:15 +0000 UTC" firstStartedPulling="2025-10-01 07:08:17.505350982 +0000 UTC m=+154.346958437" lastFinishedPulling="2025-10-01 07:09:13.117264319 +0000 UTC m=+209.958871784" observedRunningTime="2025-10-01 07:09:14.16323459 +0000 UTC m=+211.004842045" watchObservedRunningTime="2025-10-01 07:09:14.167344558 +0000 UTC m=+211.008952013" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.180964 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:09:14 crc kubenswrapper[4837]: I1001 07:09:14.190582 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.098513 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.099553 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.162495 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.417579 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.418776 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.463093 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.463348 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dcg8q" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="registry-server" containerID="cri-o://c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb" gracePeriod=2 Oct 01 07:09:15 crc kubenswrapper[4837]: I1001 07:09:15.984913 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.035095 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities\") pod \"3d2ea16a-c2ec-4f68-9c32-812598caac93\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.035189 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content\") pod \"3d2ea16a-c2ec-4f68-9c32-812598caac93\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.035232 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj62r\" (UniqueName: \"kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r\") pod \"3d2ea16a-c2ec-4f68-9c32-812598caac93\" (UID: \"3d2ea16a-c2ec-4f68-9c32-812598caac93\") " Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.036508 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities" (OuterVolumeSpecName: "utilities") pod "3d2ea16a-c2ec-4f68-9c32-812598caac93" (UID: "3d2ea16a-c2ec-4f68-9c32-812598caac93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.045885 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r" (OuterVolumeSpecName: "kube-api-access-xj62r") pod "3d2ea16a-c2ec-4f68-9c32-812598caac93" (UID: "3d2ea16a-c2ec-4f68-9c32-812598caac93"). InnerVolumeSpecName "kube-api-access-xj62r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.086804 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d2ea16a-c2ec-4f68-9c32-812598caac93" (UID: "3d2ea16a-c2ec-4f68-9c32-812598caac93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.137095 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj62r\" (UniqueName: \"kubernetes.io/projected/3d2ea16a-c2ec-4f68-9c32-812598caac93-kube-api-access-xj62r\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.137518 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.137531 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d2ea16a-c2ec-4f68-9c32-812598caac93-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.155863 4837 generic.go:334] "Generic (PLEG): container finished" podID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerID="c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb" exitCode=0 Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.156020 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dcg8q" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.156067 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerDied","Data":"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb"} Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.156142 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dcg8q" event={"ID":"3d2ea16a-c2ec-4f68-9c32-812598caac93","Type":"ContainerDied","Data":"c819a38c1981748d6da6d2ea9ac403bc172988002096da15e812d93e95056090"} Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.156171 4837 scope.go:117] "RemoveContainer" containerID="c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.192936 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.194916 4837 scope.go:117] "RemoveContainer" containerID="1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.201392 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dcg8q"] Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.213393 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.220128 4837 scope.go:117] "RemoveContainer" containerID="dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.252339 4837 scope.go:117] "RemoveContainer" containerID="c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb" Oct 01 07:09:16 crc kubenswrapper[4837]: E1001 07:09:16.253053 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb\": container with ID starting with c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb not found: ID does not exist" containerID="c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.253086 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb"} err="failed to get container status \"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb\": rpc error: code = NotFound desc = could not find container \"c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb\": container with ID starting with c0ad63fe9d96022c5cafa37730bc9b29d50ba958a3f4424f6f5e4408e9120dfb not found: ID does not exist" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.253147 4837 scope.go:117] "RemoveContainer" containerID="1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd" Oct 01 07:09:16 crc kubenswrapper[4837]: E1001 07:09:16.253801 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd\": container with ID starting with 1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd not found: ID does not exist" containerID="1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.253899 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd"} err="failed to get container status \"1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd\": rpc error: code = NotFound desc = could not find container \"1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd\": container with ID starting with 1431b7888a7fcaa5524a26e017a63d2176c6f36b9378ab038f98b781ec96c3dd not found: ID does not exist" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.253971 4837 scope.go:117] "RemoveContainer" containerID="dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0" Oct 01 07:09:16 crc kubenswrapper[4837]: E1001 07:09:16.254421 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0\": container with ID starting with dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0 not found: ID does not exist" containerID="dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.254467 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0"} err="failed to get container status \"dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0\": rpc error: code = NotFound desc = could not find container \"dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0\": container with ID starting with dc0e299ba03990e7495bc09f4bc2d3cf06a9c2494eff11d3d9bba0d0a69318c0 not found: ID does not exist" Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.455754 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q8cdq" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="registry-server" probeResult="failure" output=< Oct 01 07:09:16 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 07:09:16 crc kubenswrapper[4837]: > Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.862470 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:09:16 crc kubenswrapper[4837]: I1001 07:09:16.863352 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9k6b2" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="registry-server" containerID="cri-o://2055a4807cf966100b6adf22bca618282ee42448ddc59491151dd1e3cd335887" gracePeriod=2 Oct 01 07:09:17 crc kubenswrapper[4837]: I1001 07:09:17.832974 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" path="/var/lib/kubelet/pods/3d2ea16a-c2ec-4f68-9c32-812598caac93/volumes" Oct 01 07:09:17 crc kubenswrapper[4837]: I1001 07:09:17.861274 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:09:17 crc kubenswrapper[4837]: I1001 07:09:17.861603 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rbl6p" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="registry-server" containerID="cri-o://79184f87785f55c775fab808050fee50981575325abd70106cae97b6a7765196" gracePeriod=2 Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.171374 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerID="2055a4807cf966100b6adf22bca618282ee42448ddc59491151dd1e3cd335887" exitCode=0 Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.171423 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerDied","Data":"2055a4807cf966100b6adf22bca618282ee42448ddc59491151dd1e3cd335887"} Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.313117 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.368511 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities\") pod \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.368567 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqdjq\" (UniqueName: \"kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq\") pod \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.368609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content\") pod \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\" (UID: \"6b25ef2f-3713-4377-8e5c-c0e9ef894969\") " Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.370213 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities" (OuterVolumeSpecName: "utilities") pod "6b25ef2f-3713-4377-8e5c-c0e9ef894969" (UID: "6b25ef2f-3713-4377-8e5c-c0e9ef894969"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.377814 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq" (OuterVolumeSpecName: "kube-api-access-tqdjq") pod "6b25ef2f-3713-4377-8e5c-c0e9ef894969" (UID: "6b25ef2f-3713-4377-8e5c-c0e9ef894969"). InnerVolumeSpecName "kube-api-access-tqdjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.442536 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b25ef2f-3713-4377-8e5c-c0e9ef894969" (UID: "6b25ef2f-3713-4377-8e5c-c0e9ef894969"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.470171 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqdjq\" (UniqueName: \"kubernetes.io/projected/6b25ef2f-3713-4377-8e5c-c0e9ef894969-kube-api-access-tqdjq\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.470226 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:18 crc kubenswrapper[4837]: I1001 07:09:18.470240 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b25ef2f-3713-4377-8e5c-c0e9ef894969-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.178773 4837 generic.go:334] "Generic (PLEG): container finished" podID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerID="79184f87785f55c775fab808050fee50981575325abd70106cae97b6a7765196" exitCode=0 Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.181141 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9k6b2" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.185834 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerDied","Data":"79184f87785f55c775fab808050fee50981575325abd70106cae97b6a7765196"} Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.185958 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9k6b2" event={"ID":"6b25ef2f-3713-4377-8e5c-c0e9ef894969","Type":"ContainerDied","Data":"13b1ec5f930cc47077d6b5dc5f9d51229591e083aa5bc899378e232af149bee7"} Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.186006 4837 scope.go:117] "RemoveContainer" containerID="2055a4807cf966100b6adf22bca618282ee42448ddc59491151dd1e3cd335887" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.220946 4837 scope.go:117] "RemoveContainer" containerID="da91dfb0eb1dce808addcdacf26db4161218aebbfe6e08c117e61b0daed8891c" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.257063 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.278449 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9k6b2"] Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.278595 4837 scope.go:117] "RemoveContainer" containerID="a46d0c951ce2ec470d7480e23b9380e1bbd4e7fae5ffda14e5d463d81c2c39d6" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.412504 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.493679 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities\") pod \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.493780 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97b6t\" (UniqueName: \"kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t\") pod \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.493818 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content\") pod \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\" (UID: \"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef\") " Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.494556 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities" (OuterVolumeSpecName: "utilities") pod "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" (UID: "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.500417 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t" (OuterVolumeSpecName: "kube-api-access-97b6t") pod "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" (UID: "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef"). InnerVolumeSpecName "kube-api-access-97b6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.505961 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" (UID: "94ab3cd8-ecf2-454d-99e9-31ccc57f77ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.596001 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.596066 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97b6t\" (UniqueName: \"kubernetes.io/projected/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-kube-api-access-97b6t\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.596088 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:19 crc kubenswrapper[4837]: I1001 07:09:19.825846 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" path="/var/lib/kubelet/pods/6b25ef2f-3713-4377-8e5c-c0e9ef894969/volumes" Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.193808 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rbl6p" event={"ID":"94ab3cd8-ecf2-454d-99e9-31ccc57f77ef","Type":"ContainerDied","Data":"593b843a3651cbf81bdbb133ddc907f8cba67c61a5a831b8096ecd9868c6d40c"} Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.193897 4837 scope.go:117] "RemoveContainer" containerID="79184f87785f55c775fab808050fee50981575325abd70106cae97b6a7765196" Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.193875 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rbl6p" Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.224567 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.227736 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rbl6p"] Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.228317 4837 scope.go:117] "RemoveContainer" containerID="973a4d13cefd9888ba95a4abe0c8c5f917926a2ff97637f2b1afdc6daa4ea9f5" Oct 01 07:09:20 crc kubenswrapper[4837]: I1001 07:09:20.246189 4837 scope.go:117] "RemoveContainer" containerID="4202eb23691d964facc3aecb10fe197a2e9571cb679a2333b0a294f7f752e468" Oct 01 07:09:21 crc kubenswrapper[4837]: I1001 07:09:21.824409 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" path="/var/lib/kubelet/pods/94ab3cd8-ecf2-454d-99e9-31ccc57f77ef/volumes" Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.079592 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.080186 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.080250 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.080925 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.080992 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc" gracePeriod=600 Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.223421 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc" exitCode=0 Oct 01 07:09:23 crc kubenswrapper[4837]: I1001 07:09:23.223745 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc"} Oct 01 07:09:24 crc kubenswrapper[4837]: I1001 07:09:24.236397 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1"} Oct 01 07:09:25 crc kubenswrapper[4837]: I1001 07:09:25.468791 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:25 crc kubenswrapper[4837]: I1001 07:09:25.519663 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:29 crc kubenswrapper[4837]: I1001 07:09:29.260377 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:09:29 crc kubenswrapper[4837]: I1001 07:09:29.261008 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8cdq" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="registry-server" containerID="cri-o://114bc921be3381250b9e3e71ea03a62efae015295524a9cda3a56aa4b6e8f58a" gracePeriod=2 Oct 01 07:09:30 crc kubenswrapper[4837]: I1001 07:09:30.275795 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerID="114bc921be3381250b9e3e71ea03a62efae015295524a9cda3a56aa4b6e8f58a" exitCode=0 Oct 01 07:09:30 crc kubenswrapper[4837]: I1001 07:09:30.275880 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerDied","Data":"114bc921be3381250b9e3e71ea03a62efae015295524a9cda3a56aa4b6e8f58a"} Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.029946 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.067723 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dqrz\" (UniqueName: \"kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz\") pod \"7b94d903-fa51-451e-b95a-b251ad3302ff\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.067854 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities\") pod \"7b94d903-fa51-451e-b95a-b251ad3302ff\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.067901 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content\") pod \"7b94d903-fa51-451e-b95a-b251ad3302ff\" (UID: \"7b94d903-fa51-451e-b95a-b251ad3302ff\") " Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.069580 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities" (OuterVolumeSpecName: "utilities") pod "7b94d903-fa51-451e-b95a-b251ad3302ff" (UID: "7b94d903-fa51-451e-b95a-b251ad3302ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.089914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz" (OuterVolumeSpecName: "kube-api-access-4dqrz") pod "7b94d903-fa51-451e-b95a-b251ad3302ff" (UID: "7b94d903-fa51-451e-b95a-b251ad3302ff"). InnerVolumeSpecName "kube-api-access-4dqrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.169903 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.169949 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dqrz\" (UniqueName: \"kubernetes.io/projected/7b94d903-fa51-451e-b95a-b251ad3302ff-kube-api-access-4dqrz\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.171659 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b94d903-fa51-451e-b95a-b251ad3302ff" (UID: "7b94d903-fa51-451e-b95a-b251ad3302ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.271027 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b94d903-fa51-451e-b95a-b251ad3302ff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.285976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8cdq" event={"ID":"7b94d903-fa51-451e-b95a-b251ad3302ff","Type":"ContainerDied","Data":"099b47bd075040f668f132dfd5c131c9147da3cc5bc9fd3a924c9f687efb8496"} Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.286048 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8cdq" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.286066 4837 scope.go:117] "RemoveContainer" containerID="114bc921be3381250b9e3e71ea03a62efae015295524a9cda3a56aa4b6e8f58a" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.305895 4837 scope.go:117] "RemoveContainer" containerID="787a48808d58c3dd4d262779086efe9acacfb45bb827fc45512511cbf62c6247" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.314370 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.316845 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8cdq"] Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.336274 4837 scope.go:117] "RemoveContainer" containerID="b0e25a27de32a72378f2925c579171430bee97bd557235a90c3cba115093b649" Oct 01 07:09:31 crc kubenswrapper[4837]: I1001 07:09:31.825840 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" path="/var/lib/kubelet/pods/7b94d903-fa51-451e-b95a-b251ad3302ff/volumes" Oct 01 07:09:44 crc kubenswrapper[4837]: I1001 07:09:44.574246 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:10:09 crc kubenswrapper[4837]: I1001 07:10:09.613363 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" podUID="a6fb593e-7716-401b-b74c-de0145159433" containerName="oauth-openshift" containerID="cri-o://9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af" gracePeriod=15 Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.086149 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125371 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl"] Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125656 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125670 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125683 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125709 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125719 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548ce95e-d2b9-44f9-bc8f-c0cda36710f7" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125728 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="548ce95e-d2b9-44f9-bc8f-c0cda36710f7" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125736 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" containerName="collect-profiles" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125742 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" containerName="collect-profiles" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125750 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125757 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125765 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125771 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125780 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125785 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125794 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125799 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125809 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6fb593e-7716-401b-b74c-de0145159433" containerName="oauth-openshift" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125815 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6fb593e-7716-401b-b74c-de0145159433" containerName="oauth-openshift" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125825 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d628bd9d-69b8-48fe-962f-70986bf17703" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125832 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d628bd9d-69b8-48fe-962f-70986bf17703" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125841 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125847 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125856 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125863 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125869 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125875 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125885 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125891 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="extract-utilities" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125900 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125906 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.125921 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.125927 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="extract-content" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126022 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="548ce95e-d2b9-44f9-bc8f-c0cda36710f7" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126034 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6fb593e-7716-401b-b74c-de0145159433" containerName="oauth-openshift" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126041 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d628bd9d-69b8-48fe-962f-70986bf17703" containerName="pruner" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126052 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ab3cd8-ecf2-454d-99e9-31ccc57f77ef" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126060 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" containerName="collect-profiles" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126072 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b94d903-fa51-451e-b95a-b251ad3302ff" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126082 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d2ea16a-c2ec-4f68-9c32-812598caac93" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126091 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b25ef2f-3713-4377-8e5c-c0e9ef894969" containerName="registry-server" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.126535 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.141754 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl"] Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200391 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200471 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200572 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200598 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200634 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.200674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.201897 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202284 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.201981 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202311 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvvmf\" (UniqueName: \"kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202390 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202413 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202438 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202467 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies\") pod \"a6fb593e-7716-401b-b74c-de0145159433\" (UID: \"a6fb593e-7716-401b-b74c-de0145159433\") " Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202676 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202743 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202777 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202801 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-policies\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202892 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202922 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.202984 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-dir\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203007 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203034 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl6h5\" (UniqueName: \"kubernetes.io/projected/7c9503d8-6366-4d81-afb2-029f554bebe4-kube-api-access-rl6h5\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203062 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203092 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203115 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203140 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-session\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203183 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203198 4837 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6fb593e-7716-401b-b74c-de0145159433-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.203732 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.204477 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.204538 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.208459 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.213963 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.214925 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.215030 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.215273 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf" (OuterVolumeSpecName: "kube-api-access-kvvmf") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "kube-api-access-kvvmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.219921 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.221176 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.221187 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.221620 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a6fb593e-7716-401b-b74c-de0145159433" (UID: "a6fb593e-7716-401b-b74c-de0145159433"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.304506 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.304603 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.304681 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-dir\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.305131 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-dir\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.305239 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306217 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306681 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl6h5\" (UniqueName: \"kubernetes.io/projected/7c9503d8-6366-4d81-afb2-029f554bebe4-kube-api-access-rl6h5\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306777 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306823 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306860 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.306942 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-session\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307121 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307510 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307581 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307612 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-policies\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307770 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307811 4837 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307833 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307852 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307870 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307889 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307907 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307925 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307943 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307963 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvvmf\" (UniqueName: \"kubernetes.io/projected/a6fb593e-7716-401b-b74c-de0145159433-kube-api-access-kvvmf\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307982 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.308000 4837 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a6fb593e-7716-401b-b74c-de0145159433-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.307808 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.308811 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.310102 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7c9503d8-6366-4d81-afb2-029f554bebe4-audit-policies\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.310165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.311741 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-session\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.311969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.312266 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.312601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.312915 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.313251 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.315005 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9503d8-6366-4d81-afb2-029f554bebe4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.324137 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl6h5\" (UniqueName: \"kubernetes.io/projected/7c9503d8-6366-4d81-afb2-029f554bebe4-kube-api-access-rl6h5\") pod \"oauth-openshift-6cc7c68bbf-pwxfl\" (UID: \"7c9503d8-6366-4d81-afb2-029f554bebe4\") " pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.482647 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.552187 4837 generic.go:334] "Generic (PLEG): container finished" podID="a6fb593e-7716-401b-b74c-de0145159433" containerID="9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af" exitCode=0 Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.552271 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" event={"ID":"a6fb593e-7716-401b-b74c-de0145159433","Type":"ContainerDied","Data":"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af"} Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.552342 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" event={"ID":"a6fb593e-7716-401b-b74c-de0145159433","Type":"ContainerDied","Data":"5c5b130dcdd21472bcf5d14db54a7b11884bc9cc0c88c441b9cf3856ab288ffe"} Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.552357 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mlw9m" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.552389 4837 scope.go:117] "RemoveContainer" containerID="9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.581018 4837 scope.go:117] "RemoveContainer" containerID="9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af" Oct 01 07:10:10 crc kubenswrapper[4837]: E1001 07:10:10.582085 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af\": container with ID starting with 9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af not found: ID does not exist" containerID="9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.582187 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af"} err="failed to get container status \"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af\": rpc error: code = NotFound desc = could not find container \"9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af\": container with ID starting with 9490eded0355e80ab6c29acf0e4f6b22f2c1ed671cab2f93331511d0368cc1af not found: ID does not exist" Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.606598 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.613437 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mlw9m"] Oct 01 07:10:10 crc kubenswrapper[4837]: I1001 07:10:10.969409 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl"] Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.560648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" event={"ID":"7c9503d8-6366-4d81-afb2-029f554bebe4","Type":"ContainerStarted","Data":"8cb183398a0409c92fd151e172979261fd481cc50460398ab95ecadbe82df5a4"} Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.561190 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" event={"ID":"7c9503d8-6366-4d81-afb2-029f554bebe4","Type":"ContainerStarted","Data":"3d9291c6b62be513df74bc263c7dd22b541cbe95b2bce99e0cd58b9ec159d6fa"} Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.561209 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.588933 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" podStartSLOduration=27.588910691 podStartE2EDuration="27.588910691s" podCreationTimestamp="2025-10-01 07:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:10:11.587633979 +0000 UTC m=+268.429241464" watchObservedRunningTime="2025-10-01 07:10:11.588910691 +0000 UTC m=+268.430518156" Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.606068 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" Oct 01 07:10:11 crc kubenswrapper[4837]: I1001 07:10:11.822787 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6fb593e-7716-401b-b74c-de0145159433" path="/var/lib/kubelet/pods/a6fb593e-7716-401b-b74c-de0145159433/volumes" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.108776 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.114040 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.114381 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-njs2f" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="registry-server" containerID="cri-o://2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39" gracePeriod=30 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.120959 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jtdts" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="registry-server" containerID="cri-o://3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c" gracePeriod=30 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.138108 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.138458 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" containerID="cri-o://b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2" gracePeriod=30 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.143516 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.143864 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qhvxs" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="registry-server" containerID="cri-o://921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd" gracePeriod=30 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.153927 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.154280 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t5ql6" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="registry-server" containerID="cri-o://83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569" gracePeriod=30 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.163074 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t7lps"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.164447 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.174827 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t7lps"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.265196 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.265793 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.265996 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5bbz\" (UniqueName: \"kubernetes.io/projected/03ef9a21-4a00-49ac-9891-ad16e54b9d84-kube-api-access-p5bbz\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.367985 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.368104 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5bbz\" (UniqueName: \"kubernetes.io/projected/03ef9a21-4a00-49ac-9891-ad16e54b9d84-kube-api-access-p5bbz\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.368148 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.369818 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.376033 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/03ef9a21-4a00-49ac-9891-ad16e54b9d84-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.385517 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5bbz\" (UniqueName: \"kubernetes.io/projected/03ef9a21-4a00-49ac-9891-ad16e54b9d84-kube-api-access-p5bbz\") pod \"marketplace-operator-79b997595-t7lps\" (UID: \"03ef9a21-4a00-49ac-9891-ad16e54b9d84\") " pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.480029 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.586373 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.666216 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.671911 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.672039 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzdpr\" (UniqueName: \"kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr\") pod \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.672614 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities\") pod \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.672857 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content\") pod \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\" (UID: \"9a25e900-33fa-4d02-8ea5-377911d8e4e9\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.673395 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities" (OuterVolumeSpecName: "utilities") pod "9a25e900-33fa-4d02-8ea5-377911d8e4e9" (UID: "9a25e900-33fa-4d02-8ea5-377911d8e4e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.681529 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr" (OuterVolumeSpecName: "kube-api-access-hzdpr") pod "9a25e900-33fa-4d02-8ea5-377911d8e4e9" (UID: "9a25e900-33fa-4d02-8ea5-377911d8e4e9"). InnerVolumeSpecName "kube-api-access-hzdpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.723886 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.747045 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.765623 4837 generic.go:334] "Generic (PLEG): container finished" podID="533b6b0d-129b-4d94-b233-541162a9a58f" containerID="83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569" exitCode=0 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.765686 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerDied","Data":"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.765733 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5ql6" event={"ID":"533b6b0d-129b-4d94-b233-541162a9a58f","Type":"ContainerDied","Data":"f82cee67614c4fd010b0ae1ff6320e5fa8776cd729a25b5ca1ca61c57f5a6c3e"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.765750 4837 scope.go:117] "RemoveContainer" containerID="83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.765871 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5ql6" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.773993 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content\") pod \"534df8e2-faff-47cf-a3d4-e62505fab14a\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.774056 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8k5t\" (UniqueName: \"kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t\") pod \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775399 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhvmr\" (UniqueName: \"kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr\") pod \"533b6b0d-129b-4d94-b233-541162a9a58f\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775436 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities\") pod \"534df8e2-faff-47cf-a3d4-e62505fab14a\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775510 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca\") pod \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775537 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content\") pod \"533b6b0d-129b-4d94-b233-541162a9a58f\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775574 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t76zg\" (UniqueName: \"kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg\") pod \"534df8e2-faff-47cf-a3d4-e62505fab14a\" (UID: \"534df8e2-faff-47cf-a3d4-e62505fab14a\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775599 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities\") pod \"533b6b0d-129b-4d94-b233-541162a9a58f\" (UID: \"533b6b0d-129b-4d94-b233-541162a9a58f\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775642 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics\") pod \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\" (UID: \"f4ae7b9d-d7f0-4367-860f-81106f4b9d63\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775970 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzdpr\" (UniqueName: \"kubernetes.io/projected/9a25e900-33fa-4d02-8ea5-377911d8e4e9-kube-api-access-hzdpr\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.775989 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.776772 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f4ae7b9d-d7f0-4367-860f-81106f4b9d63" (UID: "f4ae7b9d-d7f0-4367-860f-81106f4b9d63"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.778717 4837 generic.go:334] "Generic (PLEG): container finished" podID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerID="3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c" exitCode=0 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.778867 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerDied","Data":"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.778897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtdts" event={"ID":"9a25e900-33fa-4d02-8ea5-377911d8e4e9","Type":"ContainerDied","Data":"aa8ee3f8923adac7cd5c99742d506025d47eafaaf066c6bd82c9ccde067028c8"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.778989 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtdts" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.780091 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities" (OuterVolumeSpecName: "utilities") pod "534df8e2-faff-47cf-a3d4-e62505fab14a" (UID: "534df8e2-faff-47cf-a3d4-e62505fab14a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.780723 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities" (OuterVolumeSpecName: "utilities") pod "533b6b0d-129b-4d94-b233-541162a9a58f" (UID: "533b6b0d-129b-4d94-b233-541162a9a58f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.781060 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f4ae7b9d-d7f0-4367-860f-81106f4b9d63" (UID: "f4ae7b9d-d7f0-4367-860f-81106f4b9d63"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.783398 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg" (OuterVolumeSpecName: "kube-api-access-t76zg") pod "534df8e2-faff-47cf-a3d4-e62505fab14a" (UID: "534df8e2-faff-47cf-a3d4-e62505fab14a"). InnerVolumeSpecName "kube-api-access-t76zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.784177 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t" (OuterVolumeSpecName: "kube-api-access-f8k5t") pod "f4ae7b9d-d7f0-4367-860f-81106f4b9d63" (UID: "f4ae7b9d-d7f0-4367-860f-81106f4b9d63"). InnerVolumeSpecName "kube-api-access-f8k5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.785455 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr" (OuterVolumeSpecName: "kube-api-access-hhvmr") pod "533b6b0d-129b-4d94-b233-541162a9a58f" (UID: "533b6b0d-129b-4d94-b233-541162a9a58f"). InnerVolumeSpecName "kube-api-access-hhvmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.792271 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a25e900-33fa-4d02-8ea5-377911d8e4e9" (UID: "9a25e900-33fa-4d02-8ea5-377911d8e4e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.796926 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "534df8e2-faff-47cf-a3d4-e62505fab14a" (UID: "534df8e2-faff-47cf-a3d4-e62505fab14a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.799052 4837 scope.go:117] "RemoveContainer" containerID="a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.800942 4837 generic.go:334] "Generic (PLEG): container finished" podID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerID="921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd" exitCode=0 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.801028 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qhvxs" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.801170 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerDied","Data":"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.801235 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qhvxs" event={"ID":"534df8e2-faff-47cf-a3d4-e62505fab14a","Type":"ContainerDied","Data":"4090276c9f2b12a01fdc1dc13b42f3da82f2064ab156ceab9fb953a4b934ef1d"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.806658 4837 generic.go:334] "Generic (PLEG): container finished" podID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerID="b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2" exitCode=0 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.806717 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.806739 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" event={"ID":"f4ae7b9d-d7f0-4367-860f-81106f4b9d63","Type":"ContainerDied","Data":"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.807135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j2drz" event={"ID":"f4ae7b9d-d7f0-4367-860f-81106f4b9d63","Type":"ContainerDied","Data":"ed61df8eab6b90f957df9df68cda092c623b7d42af4ad0b4e092ff30ea7103a9"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.815784 4837 generic.go:334] "Generic (PLEG): container finished" podID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerID="2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39" exitCode=0 Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.815844 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerDied","Data":"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.815888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njs2f" event={"ID":"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2","Type":"ContainerDied","Data":"2b9d2b45ebbd7c2c24a89596188edb68a51aaeb6679fb1ff4eff721b1bfb5d62"} Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.816379 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njs2f" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.825409 4837 scope.go:117] "RemoveContainer" containerID="2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.839383 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t7lps"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.842903 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.844848 4837 scope.go:117] "RemoveContainer" containerID="83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.845616 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qhvxs"] Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.845859 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569\": container with ID starting with 83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569 not found: ID does not exist" containerID="83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.845906 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569"} err="failed to get container status \"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569\": rpc error: code = NotFound desc = could not find container \"83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569\": container with ID starting with 83ecb926f1401453c295ae0f9a5527fdc5546115bc4874b9c28e0197e83d6569 not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.845969 4837 scope.go:117] "RemoveContainer" containerID="a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.846681 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c\": container with ID starting with a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c not found: ID does not exist" containerID="a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.846757 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c"} err="failed to get container status \"a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c\": rpc error: code = NotFound desc = could not find container \"a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c\": container with ID starting with a3dc05fc94b2613834f2ef445e318192130ddae0507b5bbdbcdd3ed84fc1412c not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.846780 4837 scope.go:117] "RemoveContainer" containerID="2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.846961 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f\": container with ID starting with 2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f not found: ID does not exist" containerID="2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.847015 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f"} err="failed to get container status \"2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f\": rpc error: code = NotFound desc = could not find container \"2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f\": container with ID starting with 2d7515ffcd14323413c978796b4a5449b8f0b24445764e8c84ce99d1d10b077f not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.847040 4837 scope.go:117] "RemoveContainer" containerID="3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.868288 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.869871 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j2drz"] Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.872358 4837 scope.go:117] "RemoveContainer" containerID="21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876518 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thcs2\" (UniqueName: \"kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2\") pod \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876581 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content\") pod \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876615 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities\") pod \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\" (UID: \"5adadfe2-e89d-43bf-94a4-b3cdb58efaa2\") " Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876900 4837 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876921 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t76zg\" (UniqueName: \"kubernetes.io/projected/534df8e2-faff-47cf-a3d4-e62505fab14a-kube-api-access-t76zg\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876930 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876939 4837 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876949 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876962 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8k5t\" (UniqueName: \"kubernetes.io/projected/f4ae7b9d-d7f0-4367-860f-81106f4b9d63-kube-api-access-f8k5t\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876973 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhvmr\" (UniqueName: \"kubernetes.io/projected/533b6b0d-129b-4d94-b233-541162a9a58f-kube-api-access-hhvmr\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876981 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a25e900-33fa-4d02-8ea5-377911d8e4e9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.876990 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/534df8e2-faff-47cf-a3d4-e62505fab14a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.879939 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities" (OuterVolumeSpecName: "utilities") pod "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" (UID: "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.881625 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2" (OuterVolumeSpecName: "kube-api-access-thcs2") pod "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" (UID: "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2"). InnerVolumeSpecName "kube-api-access-thcs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.893668 4837 scope.go:117] "RemoveContainer" containerID="e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.927069 4837 scope.go:117] "RemoveContainer" containerID="3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.927715 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c\": container with ID starting with 3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c not found: ID does not exist" containerID="3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.927766 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c"} err="failed to get container status \"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c\": rpc error: code = NotFound desc = could not find container \"3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c\": container with ID starting with 3d067f058df24de43e7d3aeca7145d8e045f1388cef88e90b149ff2071ed2c6c not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.927800 4837 scope.go:117] "RemoveContainer" containerID="21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.928205 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "533b6b0d-129b-4d94-b233-541162a9a58f" (UID: "533b6b0d-129b-4d94-b233-541162a9a58f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.928260 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6\": container with ID starting with 21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6 not found: ID does not exist" containerID="21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.928299 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6"} err="failed to get container status \"21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6\": rpc error: code = NotFound desc = could not find container \"21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6\": container with ID starting with 21514d36175efcd8624dd35aa65a443743f2144b53fdfb82026a8ac4b3beeae6 not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.928328 4837 scope.go:117] "RemoveContainer" containerID="e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.928634 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5\": container with ID starting with e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5 not found: ID does not exist" containerID="e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.928661 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5"} err="failed to get container status \"e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5\": rpc error: code = NotFound desc = could not find container \"e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5\": container with ID starting with e52978115e2a70d6ddee188fe8635d2859f379235f55757541b067320b8537b5 not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.928674 4837 scope.go:117] "RemoveContainer" containerID="921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.939018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" (UID: "5adadfe2-e89d-43bf-94a4-b3cdb58efaa2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.946456 4837 scope.go:117] "RemoveContainer" containerID="d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.978255 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thcs2\" (UniqueName: \"kubernetes.io/projected/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-kube-api-access-thcs2\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.978285 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.978297 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.978308 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/533b6b0d-129b-4d94-b233-541162a9a58f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.978794 4837 scope.go:117] "RemoveContainer" containerID="3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.992684 4837 scope.go:117] "RemoveContainer" containerID="921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.993451 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd\": container with ID starting with 921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd not found: ID does not exist" containerID="921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.993495 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd"} err="failed to get container status \"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd\": rpc error: code = NotFound desc = could not find container \"921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd\": container with ID starting with 921f32260bdade7cb16528d716f2a339f26a406feee1d8dcd57096dcabd656fd not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.993527 4837 scope.go:117] "RemoveContainer" containerID="d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.993943 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42\": container with ID starting with d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42 not found: ID does not exist" containerID="d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.993965 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42"} err="failed to get container status \"d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42\": rpc error: code = NotFound desc = could not find container \"d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42\": container with ID starting with d651a18d62c13df83ce8182b09718211d569b13abbbbc7ff73af4e884147cf42 not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.993987 4837 scope.go:117] "RemoveContainer" containerID="3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05" Oct 01 07:10:40 crc kubenswrapper[4837]: E1001 07:10:40.994219 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05\": container with ID starting with 3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05 not found: ID does not exist" containerID="3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.994256 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05"} err="failed to get container status \"3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05\": rpc error: code = NotFound desc = could not find container \"3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05\": container with ID starting with 3c027a346b04948af033919cfd6c243cf428253f950ccecf10cb17261227ba05 not found: ID does not exist" Oct 01 07:10:40 crc kubenswrapper[4837]: I1001 07:10:40.994280 4837 scope.go:117] "RemoveContainer" containerID="b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.021024 4837 scope.go:117] "RemoveContainer" containerID="b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2" Oct 01 07:10:41 crc kubenswrapper[4837]: E1001 07:10:41.021399 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2\": container with ID starting with b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2 not found: ID does not exist" containerID="b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.021443 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2"} err="failed to get container status \"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2\": rpc error: code = NotFound desc = could not find container \"b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2\": container with ID starting with b9e8008a90ace8e95e2ebebe71e0aa1e02bb80fe2ac2e7f527a70b81fc8f90c2 not found: ID does not exist" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.021473 4837 scope.go:117] "RemoveContainer" containerID="2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.036820 4837 scope.go:117] "RemoveContainer" containerID="37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.060259 4837 scope.go:117] "RemoveContainer" containerID="ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.089055 4837 scope.go:117] "RemoveContainer" containerID="2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39" Oct 01 07:10:41 crc kubenswrapper[4837]: E1001 07:10:41.090620 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39\": container with ID starting with 2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39 not found: ID does not exist" containerID="2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.090689 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39"} err="failed to get container status \"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39\": rpc error: code = NotFound desc = could not find container \"2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39\": container with ID starting with 2846626eed69e829830580ea84aac4399d717d56bb5261e9b3aed7c595e48d39 not found: ID does not exist" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.090741 4837 scope.go:117] "RemoveContainer" containerID="37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709" Oct 01 07:10:41 crc kubenswrapper[4837]: E1001 07:10:41.094938 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709\": container with ID starting with 37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709 not found: ID does not exist" containerID="37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.095032 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709"} err="failed to get container status \"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709\": rpc error: code = NotFound desc = could not find container \"37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709\": container with ID starting with 37a39934c2bb320e9f50bdb95953bb1dbac11db856e7540cf3bb29049556f709 not found: ID does not exist" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.100840 4837 scope.go:117] "RemoveContainer" containerID="ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f" Oct 01 07:10:41 crc kubenswrapper[4837]: E1001 07:10:41.101796 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f\": container with ID starting with ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f not found: ID does not exist" containerID="ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.101853 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f"} err="failed to get container status \"ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f\": rpc error: code = NotFound desc = could not find container \"ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f\": container with ID starting with ad4c99b377896657f191bf466c1147aed731f5ee1f1947e4802e3b7fd42ab02f not found: ID does not exist" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.104250 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.108211 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t5ql6"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.134103 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.139066 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jtdts"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.155758 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.159131 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-njs2f"] Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.836044 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" path="/var/lib/kubelet/pods/533b6b0d-129b-4d94-b233-541162a9a58f/volumes" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.837150 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" path="/var/lib/kubelet/pods/534df8e2-faff-47cf-a3d4-e62505fab14a/volumes" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.837812 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" path="/var/lib/kubelet/pods/5adadfe2-e89d-43bf-94a4-b3cdb58efaa2/volumes" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.838908 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" path="/var/lib/kubelet/pods/9a25e900-33fa-4d02-8ea5-377911d8e4e9/volumes" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.839676 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" path="/var/lib/kubelet/pods/f4ae7b9d-d7f0-4367-860f-81106f4b9d63/volumes" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.840226 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.840305 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.840330 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" event={"ID":"03ef9a21-4a00-49ac-9891-ad16e54b9d84","Type":"ContainerStarted","Data":"89f17006128fb54fd6e71c384f7ffa235bf703f3e8ccfedb52c7e792ade9f5a0"} Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.840353 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" event={"ID":"03ef9a21-4a00-49ac-9891-ad16e54b9d84","Type":"ContainerStarted","Data":"c4222c7394ff57eb1b43e5ccc7592a139e158d39f3335aa2b1dc5bd03dba8ca5"} Oct 01 07:10:41 crc kubenswrapper[4837]: I1001 07:10:41.845163 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-t7lps" podStartSLOduration=1.845147448 podStartE2EDuration="1.845147448s" podCreationTimestamp="2025-10-01 07:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:10:41.84321768 +0000 UTC m=+298.684825135" watchObservedRunningTime="2025-10-01 07:10:41.845147448 +0000 UTC m=+298.686754903" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330548 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bh2pb"] Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330823 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330839 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330851 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330860 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330875 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330882 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330892 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330897 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330907 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330913 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330924 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.330930 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="extract-content" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.330942 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331017 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331026 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331031 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331039 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331045 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331054 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331060 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331067 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331074 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="extract-utilities" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331084 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331089 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: E1001 07:10:42.331097 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331104 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331197 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5adadfe2-e89d-43bf-94a4-b3cdb58efaa2" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331209 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a25e900-33fa-4d02-8ea5-377911d8e4e9" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331216 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="534df8e2-faff-47cf-a3d4-e62505fab14a" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331223 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="533b6b0d-129b-4d94-b233-541162a9a58f" containerName="registry-server" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.331231 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ae7b9d-d7f0-4367-860f-81106f4b9d63" containerName="marketplace-operator" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.332228 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.334097 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.342837 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh2pb"] Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.400788 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-catalog-content\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.401223 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-utilities\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.401335 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljvb8\" (UniqueName: \"kubernetes.io/projected/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-kube-api-access-ljvb8\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.502312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljvb8\" (UniqueName: \"kubernetes.io/projected/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-kube-api-access-ljvb8\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.502819 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-catalog-content\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.503034 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-utilities\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.503929 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-utilities\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.504739 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-catalog-content\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.534890 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljvb8\" (UniqueName: \"kubernetes.io/projected/619e4df5-ec87-4ba4-8d5a-1f2c33f028fc-kube-api-access-ljvb8\") pod \"redhat-marketplace-bh2pb\" (UID: \"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc\") " pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.536001 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lljk9"] Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.537776 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.544855 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.547612 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lljk9"] Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.604490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-catalog-content\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.604579 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc7bd\" (UniqueName: \"kubernetes.io/projected/09d2ae22-a924-47b6-8658-eb6dbabdfa46-kube-api-access-jc7bd\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.604720 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-utilities\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.663444 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.705984 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-catalog-content\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.706031 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc7bd\" (UniqueName: \"kubernetes.io/projected/09d2ae22-a924-47b6-8658-eb6dbabdfa46-kube-api-access-jc7bd\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.706090 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-utilities\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.706490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-utilities\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.706867 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d2ae22-a924-47b6-8658-eb6dbabdfa46-catalog-content\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.739971 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc7bd\" (UniqueName: \"kubernetes.io/projected/09d2ae22-a924-47b6-8658-eb6dbabdfa46-kube-api-access-jc7bd\") pod \"redhat-operators-lljk9\" (UID: \"09d2ae22-a924-47b6-8658-eb6dbabdfa46\") " pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.871615 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:42 crc kubenswrapper[4837]: I1001 07:10:42.889021 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bh2pb"] Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.098979 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lljk9"] Oct 01 07:10:43 crc kubenswrapper[4837]: W1001 07:10:43.174981 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09d2ae22_a924_47b6_8658_eb6dbabdfa46.slice/crio-d6e0f24c56a7addfb4023fd30c1a55fba651c7c54f34d74dc77b544933b7cbdb WatchSource:0}: Error finding container d6e0f24c56a7addfb4023fd30c1a55fba651c7c54f34d74dc77b544933b7cbdb: Status 404 returned error can't find the container with id d6e0f24c56a7addfb4023fd30c1a55fba651c7c54f34d74dc77b544933b7cbdb Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.847800 4837 generic.go:334] "Generic (PLEG): container finished" podID="619e4df5-ec87-4ba4-8d5a-1f2c33f028fc" containerID="5d79639345c37c061795d964216414f5c6bf9c13f63f9732ba886f8784df6700" exitCode=0 Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.847935 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh2pb" event={"ID":"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc","Type":"ContainerDied","Data":"5d79639345c37c061795d964216414f5c6bf9c13f63f9732ba886f8784df6700"} Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.848017 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh2pb" event={"ID":"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc","Type":"ContainerStarted","Data":"327382bc4841a9f1b2ccea6265e9cd2d230a345df212849687dae9eaa4d7e6f6"} Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.857871 4837 generic.go:334] "Generic (PLEG): container finished" podID="09d2ae22-a924-47b6-8658-eb6dbabdfa46" containerID="53afb824a96c1fc31ceb6d6bd6c0d5b84f088a48acf25d548649f09aa1292b16" exitCode=0 Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.859074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljk9" event={"ID":"09d2ae22-a924-47b6-8658-eb6dbabdfa46","Type":"ContainerDied","Data":"53afb824a96c1fc31ceb6d6bd6c0d5b84f088a48acf25d548649f09aa1292b16"} Oct 01 07:10:43 crc kubenswrapper[4837]: I1001 07:10:43.859132 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljk9" event={"ID":"09d2ae22-a924-47b6-8658-eb6dbabdfa46","Type":"ContainerStarted","Data":"d6e0f24c56a7addfb4023fd30c1a55fba651c7c54f34d74dc77b544933b7cbdb"} Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.736870 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.739413 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.741990 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.745559 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.836406 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.836899 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlv7d\" (UniqueName: \"kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.836957 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.866762 4837 generic.go:334] "Generic (PLEG): container finished" podID="619e4df5-ec87-4ba4-8d5a-1f2c33f028fc" containerID="f4337b6a861a241c2094f438c19776ff1892c49090fae8bfe686d74793a21f01" exitCode=0 Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.866821 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh2pb" event={"ID":"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc","Type":"ContainerDied","Data":"f4337b6a861a241c2094f438c19776ff1892c49090fae8bfe686d74793a21f01"} Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.926449 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gdh98"] Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.927709 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.929844 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.938787 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdh98"] Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939041 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97l97\" (UniqueName: \"kubernetes.io/projected/22813b4c-26e4-41f6-ae6a-61a735db8ec5-kube-api-access-97l97\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939125 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlv7d\" (UniqueName: \"kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-catalog-content\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939231 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939286 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.939324 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-utilities\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.940781 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.941063 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:44 crc kubenswrapper[4837]: I1001 07:10:44.973815 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlv7d\" (UniqueName: \"kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d\") pod \"community-operators-pvrtj\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.040764 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-utilities\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.040848 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97l97\" (UniqueName: \"kubernetes.io/projected/22813b4c-26e4-41f6-ae6a-61a735db8ec5-kube-api-access-97l97\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.040899 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-catalog-content\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.041465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-utilities\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.041616 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22813b4c-26e4-41f6-ae6a-61a735db8ec5-catalog-content\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.060349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97l97\" (UniqueName: \"kubernetes.io/projected/22813b4c-26e4-41f6-ae6a-61a735db8ec5-kube-api-access-97l97\") pod \"certified-operators-gdh98\" (UID: \"22813b4c-26e4-41f6-ae6a-61a735db8ec5\") " pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.062819 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.267935 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.319272 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:10:45 crc kubenswrapper[4837]: W1001 07:10:45.321412 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9872972_efd5_469f_8908_e4cbe13c2565.slice/crio-5ce507ddb82c0a22eb33d5b90bfcdf5114857a3eeca3d1f0a9a6e74448bef73d WatchSource:0}: Error finding container 5ce507ddb82c0a22eb33d5b90bfcdf5114857a3eeca3d1f0a9a6e74448bef73d: Status 404 returned error can't find the container with id 5ce507ddb82c0a22eb33d5b90bfcdf5114857a3eeca3d1f0a9a6e74448bef73d Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.523303 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdh98"] Oct 01 07:10:45 crc kubenswrapper[4837]: W1001 07:10:45.573213 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22813b4c_26e4_41f6_ae6a_61a735db8ec5.slice/crio-240048458a5a6665359fb9592208098c0b04c821dd67667ccf789323d6ab6f19 WatchSource:0}: Error finding container 240048458a5a6665359fb9592208098c0b04c821dd67667ccf789323d6ab6f19: Status 404 returned error can't find the container with id 240048458a5a6665359fb9592208098c0b04c821dd67667ccf789323d6ab6f19 Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.876486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bh2pb" event={"ID":"619e4df5-ec87-4ba4-8d5a-1f2c33f028fc","Type":"ContainerStarted","Data":"eabe9f8a86d72dc3fe7b7b169db8c7b0fb513857a3249b98ed4c574dd58a470e"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.879975 4837 generic.go:334] "Generic (PLEG): container finished" podID="22813b4c-26e4-41f6-ae6a-61a735db8ec5" containerID="bdf2638bb753fbf76534f9e4e8aeac671f2c7254b93dbf7a5b1a28056dc8f9bc" exitCode=0 Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.880068 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdh98" event={"ID":"22813b4c-26e4-41f6-ae6a-61a735db8ec5","Type":"ContainerDied","Data":"bdf2638bb753fbf76534f9e4e8aeac671f2c7254b93dbf7a5b1a28056dc8f9bc"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.880118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdh98" event={"ID":"22813b4c-26e4-41f6-ae6a-61a735db8ec5","Type":"ContainerStarted","Data":"240048458a5a6665359fb9592208098c0b04c821dd67667ccf789323d6ab6f19"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.892630 4837 generic.go:334] "Generic (PLEG): container finished" podID="09d2ae22-a924-47b6-8658-eb6dbabdfa46" containerID="c91f75d0e4d6e1fc79f4c8c353340c492a8866ce2baebb28616de4d353191152" exitCode=0 Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.892778 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljk9" event={"ID":"09d2ae22-a924-47b6-8658-eb6dbabdfa46","Type":"ContainerDied","Data":"c91f75d0e4d6e1fc79f4c8c353340c492a8866ce2baebb28616de4d353191152"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.895197 4837 generic.go:334] "Generic (PLEG): container finished" podID="c9872972-efd5-469f-8908-e4cbe13c2565" containerID="39cad00979359b607b90a8a63c76d1a0aba83f567bdf78eb02b42c91eb4924fb" exitCode=0 Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.895234 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerDied","Data":"39cad00979359b607b90a8a63c76d1a0aba83f567bdf78eb02b42c91eb4924fb"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.895255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerStarted","Data":"5ce507ddb82c0a22eb33d5b90bfcdf5114857a3eeca3d1f0a9a6e74448bef73d"} Oct 01 07:10:45 crc kubenswrapper[4837]: I1001 07:10:45.900071 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bh2pb" podStartSLOduration=2.433388082 podStartE2EDuration="3.900057331s" podCreationTimestamp="2025-10-01 07:10:42 +0000 UTC" firstStartedPulling="2025-10-01 07:10:43.855488712 +0000 UTC m=+300.697096167" lastFinishedPulling="2025-10-01 07:10:45.322157961 +0000 UTC m=+302.163765416" observedRunningTime="2025-10-01 07:10:45.89718989 +0000 UTC m=+302.738797355" watchObservedRunningTime="2025-10-01 07:10:45.900057331 +0000 UTC m=+302.741664786" Oct 01 07:10:46 crc kubenswrapper[4837]: I1001 07:10:46.901900 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerStarted","Data":"3a9f765b114ddae09b09195662e26b966f86c5f6a2d13f9251e811474af89094"} Oct 01 07:10:46 crc kubenswrapper[4837]: I1001 07:10:46.906239 4837 generic.go:334] "Generic (PLEG): container finished" podID="22813b4c-26e4-41f6-ae6a-61a735db8ec5" containerID="be7eb10259f2e43f317fb0153ab9e4edb659cb3b8d476d58d1d9ef8a39bc28eb" exitCode=0 Oct 01 07:10:46 crc kubenswrapper[4837]: I1001 07:10:46.906297 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdh98" event={"ID":"22813b4c-26e4-41f6-ae6a-61a735db8ec5","Type":"ContainerDied","Data":"be7eb10259f2e43f317fb0153ab9e4edb659cb3b8d476d58d1d9ef8a39bc28eb"} Oct 01 07:10:46 crc kubenswrapper[4837]: I1001 07:10:46.930425 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljk9" event={"ID":"09d2ae22-a924-47b6-8658-eb6dbabdfa46","Type":"ContainerStarted","Data":"fde3be1ab55fdad2fc7548453268df3b36d64324eaa1cae4a06a5e15a9f48cfc"} Oct 01 07:10:46 crc kubenswrapper[4837]: E1001 07:10:46.959942 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22813b4c_26e4_41f6_ae6a_61a735db8ec5.slice/crio-conmon-be7eb10259f2e43f317fb0153ab9e4edb659cb3b8d476d58d1d9ef8a39bc28eb.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:10:46 crc kubenswrapper[4837]: I1001 07:10:46.976935 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lljk9" podStartSLOduration=2.359322216 podStartE2EDuration="4.976913934s" podCreationTimestamp="2025-10-01 07:10:42 +0000 UTC" firstStartedPulling="2025-10-01 07:10:43.863426508 +0000 UTC m=+300.705033963" lastFinishedPulling="2025-10-01 07:10:46.481018236 +0000 UTC m=+303.322625681" observedRunningTime="2025-10-01 07:10:46.974078644 +0000 UTC m=+303.815686109" watchObservedRunningTime="2025-10-01 07:10:46.976913934 +0000 UTC m=+303.818521389" Oct 01 07:10:47 crc kubenswrapper[4837]: I1001 07:10:47.939564 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdh98" event={"ID":"22813b4c-26e4-41f6-ae6a-61a735db8ec5","Type":"ContainerStarted","Data":"0aaa3e9e8c35563eff63de1d64198448b2df3a118a04fdde1827c7834e93ace5"} Oct 01 07:10:47 crc kubenswrapper[4837]: I1001 07:10:47.945383 4837 generic.go:334] "Generic (PLEG): container finished" podID="c9872972-efd5-469f-8908-e4cbe13c2565" containerID="3a9f765b114ddae09b09195662e26b966f86c5f6a2d13f9251e811474af89094" exitCode=0 Oct 01 07:10:47 crc kubenswrapper[4837]: I1001 07:10:47.945499 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerDied","Data":"3a9f765b114ddae09b09195662e26b966f86c5f6a2d13f9251e811474af89094"} Oct 01 07:10:47 crc kubenswrapper[4837]: I1001 07:10:47.961410 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gdh98" podStartSLOduration=2.540058781 podStartE2EDuration="3.961386432s" podCreationTimestamp="2025-10-01 07:10:44 +0000 UTC" firstStartedPulling="2025-10-01 07:10:45.88258677 +0000 UTC m=+302.724194225" lastFinishedPulling="2025-10-01 07:10:47.303914411 +0000 UTC m=+304.145521876" observedRunningTime="2025-10-01 07:10:47.958547732 +0000 UTC m=+304.800155187" watchObservedRunningTime="2025-10-01 07:10:47.961386432 +0000 UTC m=+304.802993887" Oct 01 07:10:48 crc kubenswrapper[4837]: I1001 07:10:48.954120 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerStarted","Data":"f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c"} Oct 01 07:10:48 crc kubenswrapper[4837]: I1001 07:10:48.977501 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pvrtj" podStartSLOduration=2.533508049 podStartE2EDuration="4.977482128s" podCreationTimestamp="2025-10-01 07:10:44 +0000 UTC" firstStartedPulling="2025-10-01 07:10:45.901540717 +0000 UTC m=+302.743148182" lastFinishedPulling="2025-10-01 07:10:48.345514806 +0000 UTC m=+305.187122261" observedRunningTime="2025-10-01 07:10:48.97470579 +0000 UTC m=+305.816313255" watchObservedRunningTime="2025-10-01 07:10:48.977482128 +0000 UTC m=+305.819089603" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.664165 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.664908 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.732718 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.871934 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.872021 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:52 crc kubenswrapper[4837]: I1001 07:10:52.939972 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:53 crc kubenswrapper[4837]: I1001 07:10:53.018802 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lljk9" Oct 01 07:10:53 crc kubenswrapper[4837]: I1001 07:10:53.019953 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bh2pb" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.064201 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.066476 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.131572 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.268778 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.269167 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:55 crc kubenswrapper[4837]: I1001 07:10:55.321661 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:10:56 crc kubenswrapper[4837]: I1001 07:10:56.102236 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:10:56 crc kubenswrapper[4837]: I1001 07:10:56.105345 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gdh98" Oct 01 07:11:23 crc kubenswrapper[4837]: I1001 07:11:23.079952 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:11:23 crc kubenswrapper[4837]: I1001 07:11:23.080649 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:11:53 crc kubenswrapper[4837]: I1001 07:11:53.080230 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:11:53 crc kubenswrapper[4837]: I1001 07:11:53.081043 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.079725 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.080619 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.080732 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.081988 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.082101 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1" gracePeriod=600 Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.603608 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1" exitCode=0 Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.603662 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1"} Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.604049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b"} Oct 01 07:12:23 crc kubenswrapper[4837]: I1001 07:12:23.604078 4837 scope.go:117] "RemoveContainer" containerID="01c192230600a79d5a3c87007662aecc10a76bf91a29b9eb95772880cafc38dc" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.705813 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j6jvg"] Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.707551 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.729077 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j6jvg"] Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809767 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-tls\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62bxw\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-kube-api-access-62bxw\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809892 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809935 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-bound-sa-token\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809961 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-certificates\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.809987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-trusted-ca\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.810079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.810122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.847452 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.910876 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62bxw\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-kube-api-access-62bxw\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.910920 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.910939 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-bound-sa-token\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.910960 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-certificates\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.910978 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-trusted-ca\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.911002 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.911057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-tls\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.911754 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.912610 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-trusted-ca\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.912834 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-certificates\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.919229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.919382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-registry-tls\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.930664 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62bxw\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-kube-api-access-62bxw\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:50 crc kubenswrapper[4837]: I1001 07:13:50.931425 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01a18e08-d4f2-4e95-99c4-f44bf8dff5ae-bound-sa-token\") pod \"image-registry-66df7c8f76-j6jvg\" (UID: \"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae\") " pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:51 crc kubenswrapper[4837]: I1001 07:13:51.028664 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:51 crc kubenswrapper[4837]: I1001 07:13:51.328670 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j6jvg"] Oct 01 07:13:52 crc kubenswrapper[4837]: I1001 07:13:52.262882 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" event={"ID":"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae","Type":"ContainerStarted","Data":"650c1e15c0e1ce525fdddf589b24278cfa843e3a5fe2fbfecdfa50bfc6b07e3d"} Oct 01 07:13:52 crc kubenswrapper[4837]: I1001 07:13:52.262958 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" event={"ID":"01a18e08-d4f2-4e95-99c4-f44bf8dff5ae","Type":"ContainerStarted","Data":"da68c1568c8a33e1e4aad2a117beb57339b103067fb9307d4d38abff506f7e7d"} Oct 01 07:13:52 crc kubenswrapper[4837]: I1001 07:13:52.263116 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:13:52 crc kubenswrapper[4837]: I1001 07:13:52.298399 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" podStartSLOduration=2.298360269 podStartE2EDuration="2.298360269s" podCreationTimestamp="2025-10-01 07:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:13:52.294595986 +0000 UTC m=+489.136203471" watchObservedRunningTime="2025-10-01 07:13:52.298360269 +0000 UTC m=+489.139967774" Oct 01 07:14:11 crc kubenswrapper[4837]: I1001 07:14:11.037817 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-j6jvg" Oct 01 07:14:11 crc kubenswrapper[4837]: I1001 07:14:11.118817 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:14:23 crc kubenswrapper[4837]: I1001 07:14:23.079478 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:14:23 crc kubenswrapper[4837]: I1001 07:14:23.080240 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.170551 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" podUID="56966f17-7502-47e9-a482-0e9f3fc63b0d" containerName="registry" containerID="cri-o://5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335" gracePeriod=30 Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.545802 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.581849 4837 generic.go:334] "Generic (PLEG): container finished" podID="56966f17-7502-47e9-a482-0e9f3fc63b0d" containerID="5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335" exitCode=0 Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.581936 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" event={"ID":"56966f17-7502-47e9-a482-0e9f3fc63b0d","Type":"ContainerDied","Data":"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335"} Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.581976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" event={"ID":"56966f17-7502-47e9-a482-0e9f3fc63b0d","Type":"ContainerDied","Data":"3a1023e97dc05b25448bf7ca555f2bde5e2f6a05d89c6218b43fd3347941aad0"} Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.582007 4837 scope.go:117] "RemoveContainer" containerID="5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.582271 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vxkzq" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.611244 4837 scope.go:117] "RemoveContainer" containerID="5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335" Oct 01 07:14:36 crc kubenswrapper[4837]: E1001 07:14:36.612055 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335\": container with ID starting with 5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335 not found: ID does not exist" containerID="5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.612095 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335"} err="failed to get container status \"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335\": rpc error: code = NotFound desc = could not find container \"5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335\": container with ID starting with 5ad1ab339a4cfe8b0470cc6a841a69b486945b123ae6205abcad55e465782335 not found: ID does not exist" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636505 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636599 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtzng\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636624 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636719 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636779 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.636812 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.637720 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.637764 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls\") pod \"56966f17-7502-47e9-a482-0e9f3fc63b0d\" (UID: \"56966f17-7502-47e9-a482-0e9f3fc63b0d\") " Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.638323 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.643584 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.648687 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.648860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng" (OuterVolumeSpecName: "kube-api-access-rtzng") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "kube-api-access-rtzng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.648965 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.657374 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.664474 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.667443 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "56966f17-7502-47e9-a482-0e9f3fc63b0d" (UID: "56966f17-7502-47e9-a482-0e9f3fc63b0d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739108 4837 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739413 4837 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56966f17-7502-47e9-a482-0e9f3fc63b0d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739537 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739634 4837 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739771 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtzng\" (UniqueName: \"kubernetes.io/projected/56966f17-7502-47e9-a482-0e9f3fc63b0d-kube-api-access-rtzng\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.739892 4837 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56966f17-7502-47e9-a482-0e9f3fc63b0d-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.740032 4837 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56966f17-7502-47e9-a482-0e9f3fc63b0d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.928326 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:14:36 crc kubenswrapper[4837]: I1001 07:14:36.937215 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vxkzq"] Oct 01 07:14:37 crc kubenswrapper[4837]: I1001 07:14:37.823737 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56966f17-7502-47e9-a482-0e9f3fc63b0d" path="/var/lib/kubelet/pods/56966f17-7502-47e9-a482-0e9f3fc63b0d/volumes" Oct 01 07:14:53 crc kubenswrapper[4837]: I1001 07:14:53.079612 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:14:53 crc kubenswrapper[4837]: I1001 07:14:53.080309 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.161586 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm"] Oct 01 07:15:00 crc kubenswrapper[4837]: E1001 07:15:00.162788 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56966f17-7502-47e9-a482-0e9f3fc63b0d" containerName="registry" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.162810 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="56966f17-7502-47e9-a482-0e9f3fc63b0d" containerName="registry" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.162998 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="56966f17-7502-47e9-a482-0e9f3fc63b0d" containerName="registry" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.163873 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.167435 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.168820 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm"] Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.175872 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.361323 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.361459 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.361531 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5tr9\" (UniqueName: \"kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.462853 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.462956 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.463010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5tr9\" (UniqueName: \"kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.464751 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.475088 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.496726 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5tr9\" (UniqueName: \"kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9\") pod \"collect-profiles-29321715-pmfnm\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:00 crc kubenswrapper[4837]: I1001 07:15:00.794763 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:01 crc kubenswrapper[4837]: I1001 07:15:01.043393 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm"] Oct 01 07:15:01 crc kubenswrapper[4837]: I1001 07:15:01.757748 4837 generic.go:334] "Generic (PLEG): container finished" podID="d355c4fd-132f-45da-a9cd-94814064f59a" containerID="ea9a586dd5ca02047896b806fc47eaec453fedcf10a74898cb12ba92523202ce" exitCode=0 Oct 01 07:15:01 crc kubenswrapper[4837]: I1001 07:15:01.757822 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" event={"ID":"d355c4fd-132f-45da-a9cd-94814064f59a","Type":"ContainerDied","Data":"ea9a586dd5ca02047896b806fc47eaec453fedcf10a74898cb12ba92523202ce"} Oct 01 07:15:01 crc kubenswrapper[4837]: I1001 07:15:01.757875 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" event={"ID":"d355c4fd-132f-45da-a9cd-94814064f59a","Type":"ContainerStarted","Data":"f326f1500f5a00476749d4f697f57e22c02d1ad22c9a0e867a48fc687ca8b05c"} Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.092685 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.210788 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume\") pod \"d355c4fd-132f-45da-a9cd-94814064f59a\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.210969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5tr9\" (UniqueName: \"kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9\") pod \"d355c4fd-132f-45da-a9cd-94814064f59a\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.210997 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume\") pod \"d355c4fd-132f-45da-a9cd-94814064f59a\" (UID: \"d355c4fd-132f-45da-a9cd-94814064f59a\") " Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.211684 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume" (OuterVolumeSpecName: "config-volume") pod "d355c4fd-132f-45da-a9cd-94814064f59a" (UID: "d355c4fd-132f-45da-a9cd-94814064f59a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.220722 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d355c4fd-132f-45da-a9cd-94814064f59a" (UID: "d355c4fd-132f-45da-a9cd-94814064f59a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.220988 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9" (OuterVolumeSpecName: "kube-api-access-l5tr9") pod "d355c4fd-132f-45da-a9cd-94814064f59a" (UID: "d355c4fd-132f-45da-a9cd-94814064f59a"). InnerVolumeSpecName "kube-api-access-l5tr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.312567 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5tr9\" (UniqueName: \"kubernetes.io/projected/d355c4fd-132f-45da-a9cd-94814064f59a-kube-api-access-l5tr9\") on node \"crc\" DevicePath \"\"" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.312631 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d355c4fd-132f-45da-a9cd-94814064f59a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.312655 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d355c4fd-132f-45da-a9cd-94814064f59a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.775199 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" event={"ID":"d355c4fd-132f-45da-a9cd-94814064f59a","Type":"ContainerDied","Data":"f326f1500f5a00476749d4f697f57e22c02d1ad22c9a0e867a48fc687ca8b05c"} Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.775545 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f326f1500f5a00476749d4f697f57e22c02d1ad22c9a0e867a48fc687ca8b05c" Oct 01 07:15:03 crc kubenswrapper[4837]: I1001 07:15:03.775279 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm" Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.080316 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.081207 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.081278 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.082185 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.082285 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b" gracePeriod=600 Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.928727 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b"} Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.928675 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b" exitCode=0 Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.929632 4837 scope.go:117] "RemoveContainer" containerID="fd2433bafc926dc873e09e84c80232adc4a7ded8aedf5ab47db180dc9444dfb1" Oct 01 07:15:23 crc kubenswrapper[4837]: I1001 07:15:23.929747 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05"} Oct 01 07:17:23 crc kubenswrapper[4837]: I1001 07:17:23.079407 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:17:23 crc kubenswrapper[4837]: I1001 07:17:23.080395 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:17:53 crc kubenswrapper[4837]: I1001 07:17:53.079948 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:17:53 crc kubenswrapper[4837]: I1001 07:17:53.080528 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:18:11 crc kubenswrapper[4837]: I1001 07:18:11.649229 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:18:11 crc kubenswrapper[4837]: I1001 07:18:11.652272 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" containerName="controller-manager" containerID="cri-o://d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7" gracePeriod=30 Oct 01 07:18:11 crc kubenswrapper[4837]: I1001 07:18:11.736530 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:18:11 crc kubenswrapper[4837]: I1001 07:18:11.736786 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerName="route-controller-manager" containerID="cri-o://e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7" gracePeriod=30 Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.085742 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.113332 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.162349 4837 generic.go:334] "Generic (PLEG): container finished" podID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerID="e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7" exitCode=0 Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.162400 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.162418 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" event={"ID":"6042d50d-4f20-47d8-a9d8-74c8424aff37","Type":"ContainerDied","Data":"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7"} Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.162473 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh" event={"ID":"6042d50d-4f20-47d8-a9d8-74c8424aff37","Type":"ContainerDied","Data":"52265347b20baea6d5ef4077ab7fdebe50599ffa8dcd6321298420074d19d8e3"} Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.162498 4837 scope.go:117] "RemoveContainer" containerID="e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.165325 4837 generic.go:334] "Generic (PLEG): container finished" podID="25fe367b-044a-4d96-9658-eaad1a16582a" containerID="d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7" exitCode=0 Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.165358 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" event={"ID":"25fe367b-044a-4d96-9658-eaad1a16582a","Type":"ContainerDied","Data":"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7"} Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.165383 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" event={"ID":"25fe367b-044a-4d96-9658-eaad1a16582a","Type":"ContainerDied","Data":"fa4dd93a848f180c7a538f89b97892bd2fe001d7fe603371059da95437f4fee9"} Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.165431 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4cx5v" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.181543 4837 scope.go:117] "RemoveContainer" containerID="e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7" Oct 01 07:18:12 crc kubenswrapper[4837]: E1001 07:18:12.182288 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7\": container with ID starting with e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7 not found: ID does not exist" containerID="e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.182326 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7"} err="failed to get container status \"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7\": rpc error: code = NotFound desc = could not find container \"e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7\": container with ID starting with e9acc5aa8bb111e08fa54f8310808b356ddba33ab585b724a95788997af298e7 not found: ID does not exist" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.182360 4837 scope.go:117] "RemoveContainer" containerID="d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.195673 4837 scope.go:117] "RemoveContainer" containerID="d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7" Oct 01 07:18:12 crc kubenswrapper[4837]: E1001 07:18:12.196034 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7\": container with ID starting with d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7 not found: ID does not exist" containerID="d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.196060 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7"} err="failed to get container status \"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7\": rpc error: code = NotFound desc = could not find container \"d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7\": container with ID starting with d1bdee7a44d1164c5e5c633a821fd2dc9c917181f4de255cd8489b2c24c380d7 not found: ID does not exist" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200652 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca\") pod \"6042d50d-4f20-47d8-a9d8-74c8424aff37\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200731 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvftf\" (UniqueName: \"kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf\") pod \"25fe367b-044a-4d96-9658-eaad1a16582a\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200762 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles\") pod \"25fe367b-044a-4d96-9658-eaad1a16582a\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200790 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config\") pod \"6042d50d-4f20-47d8-a9d8-74c8424aff37\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200830 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert\") pod \"6042d50d-4f20-47d8-a9d8-74c8424aff37\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200879 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert\") pod \"25fe367b-044a-4d96-9658-eaad1a16582a\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.200905 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqk4x\" (UniqueName: \"kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x\") pod \"6042d50d-4f20-47d8-a9d8-74c8424aff37\" (UID: \"6042d50d-4f20-47d8-a9d8-74c8424aff37\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.201540 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config\") pod \"25fe367b-044a-4d96-9658-eaad1a16582a\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.201548 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "25fe367b-044a-4d96-9658-eaad1a16582a" (UID: "25fe367b-044a-4d96-9658-eaad1a16582a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.202034 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca\") pod \"25fe367b-044a-4d96-9658-eaad1a16582a\" (UID: \"25fe367b-044a-4d96-9658-eaad1a16582a\") " Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.202034 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config" (OuterVolumeSpecName: "config") pod "6042d50d-4f20-47d8-a9d8-74c8424aff37" (UID: "6042d50d-4f20-47d8-a9d8-74c8424aff37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.202143 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca" (OuterVolumeSpecName: "client-ca") pod "6042d50d-4f20-47d8-a9d8-74c8424aff37" (UID: "6042d50d-4f20-47d8-a9d8-74c8424aff37"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.202332 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config" (OuterVolumeSpecName: "config") pod "25fe367b-044a-4d96-9658-eaad1a16582a" (UID: "25fe367b-044a-4d96-9658-eaad1a16582a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.202440 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca" (OuterVolumeSpecName: "client-ca") pod "25fe367b-044a-4d96-9658-eaad1a16582a" (UID: "25fe367b-044a-4d96-9658-eaad1a16582a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207142 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207222 4837 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207242 4837 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207253 4837 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25fe367b-044a-4d96-9658-eaad1a16582a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207274 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6042d50d-4f20-47d8-a9d8-74c8424aff37-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207432 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x" (OuterVolumeSpecName: "kube-api-access-wqk4x") pod "6042d50d-4f20-47d8-a9d8-74c8424aff37" (UID: "6042d50d-4f20-47d8-a9d8-74c8424aff37"). InnerVolumeSpecName "kube-api-access-wqk4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.207720 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6042d50d-4f20-47d8-a9d8-74c8424aff37" (UID: "6042d50d-4f20-47d8-a9d8-74c8424aff37"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.208027 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "25fe367b-044a-4d96-9658-eaad1a16582a" (UID: "25fe367b-044a-4d96-9658-eaad1a16582a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.208364 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf" (OuterVolumeSpecName: "kube-api-access-wvftf") pod "25fe367b-044a-4d96-9658-eaad1a16582a" (UID: "25fe367b-044a-4d96-9658-eaad1a16582a"). InnerVolumeSpecName "kube-api-access-wvftf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.311993 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25fe367b-044a-4d96-9658-eaad1a16582a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.312030 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqk4x\" (UniqueName: \"kubernetes.io/projected/6042d50d-4f20-47d8-a9d8-74c8424aff37-kube-api-access-wqk4x\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.312043 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvftf\" (UniqueName: \"kubernetes.io/projected/25fe367b-044a-4d96-9658-eaad1a16582a-kube-api-access-wvftf\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.312051 4837 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6042d50d-4f20-47d8-a9d8-74c8424aff37-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.491534 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.496518 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t8vmh"] Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.510587 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:18:12 crc kubenswrapper[4837]: I1001 07:18:12.515932 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4cx5v"] Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770356 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-675ff74c79-5cg2k"] Oct 01 07:18:13 crc kubenswrapper[4837]: E1001 07:18:13.770629 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d355c4fd-132f-45da-a9cd-94814064f59a" containerName="collect-profiles" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770644 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d355c4fd-132f-45da-a9cd-94814064f59a" containerName="collect-profiles" Oct 01 07:18:13 crc kubenswrapper[4837]: E1001 07:18:13.770667 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" containerName="controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770674 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" containerName="controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: E1001 07:18:13.770708 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerName="route-controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770716 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerName="route-controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770814 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d355c4fd-132f-45da-a9cd-94814064f59a" containerName="collect-profiles" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770834 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" containerName="controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.770844 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" containerName="route-controller-manager" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.771277 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.776141 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp"] Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.777523 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.778681 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.778799 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.778962 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.780493 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.780684 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.780975 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.780984 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.781476 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.781728 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.781994 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.782067 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.782281 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.786840 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.796904 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-675ff74c79-5cg2k"] Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.800167 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp"] Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.845556 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25fe367b-044a-4d96-9658-eaad1a16582a" path="/var/lib/kubelet/pods/25fe367b-044a-4d96-9658-eaad1a16582a/volumes" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.846815 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6042d50d-4f20-47d8-a9d8-74c8424aff37" path="/var/lib/kubelet/pods/6042d50d-4f20-47d8-a9d8-74c8424aff37/volumes" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-config\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935116 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-serving-cert\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935135 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-client-ca\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935153 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pbj\" (UniqueName: \"kubernetes.io/projected/8cb8d05a-e31a-4567-be65-b33c94f7c67f-kube-api-access-78pbj\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935176 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-proxy-ca-bundles\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935306 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp7df\" (UniqueName: \"kubernetes.io/projected/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-kube-api-access-qp7df\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935404 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cb8d05a-e31a-4567-be65-b33c94f7c67f-serving-cert\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935594 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-client-ca\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:13 crc kubenswrapper[4837]: I1001 07:18:13.935766 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-config\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037214 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-client-ca\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037295 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-config\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037343 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-config\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037377 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-serving-cert\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037417 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-client-ca\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037450 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pbj\" (UniqueName: \"kubernetes.io/projected/8cb8d05a-e31a-4567-be65-b33c94f7c67f-kube-api-access-78pbj\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037491 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-proxy-ca-bundles\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037538 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp7df\" (UniqueName: \"kubernetes.io/projected/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-kube-api-access-qp7df\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.037582 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cb8d05a-e31a-4567-be65-b33c94f7c67f-serving-cert\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.039051 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-proxy-ca-bundles\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.039729 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-client-ca\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.039956 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-client-ca\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.040368 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cb8d05a-e31a-4567-be65-b33c94f7c67f-config\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.040935 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-config\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.047398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-serving-cert\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.047475 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cb8d05a-e31a-4567-be65-b33c94f7c67f-serving-cert\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.064797 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pbj\" (UniqueName: \"kubernetes.io/projected/8cb8d05a-e31a-4567-be65-b33c94f7c67f-kube-api-access-78pbj\") pod \"route-controller-manager-7bdd8c968c-h24zp\" (UID: \"8cb8d05a-e31a-4567-be65-b33c94f7c67f\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.071239 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp7df\" (UniqueName: \"kubernetes.io/projected/d80beefa-ae96-4167-9bbe-6d2b0cfd8605-kube-api-access-qp7df\") pod \"controller-manager-675ff74c79-5cg2k\" (UID: \"d80beefa-ae96-4167-9bbe-6d2b0cfd8605\") " pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.093094 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.102635 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.333151 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-675ff74c79-5cg2k"] Oct 01 07:18:14 crc kubenswrapper[4837]: I1001 07:18:14.378937 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp"] Oct 01 07:18:14 crc kubenswrapper[4837]: W1001 07:18:14.384300 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cb8d05a_e31a_4567_be65_b33c94f7c67f.slice/crio-36620aaebd7df7c81208ba26284d4ffc384d617e2cdf2186246988edc3130d5b WatchSource:0}: Error finding container 36620aaebd7df7c81208ba26284d4ffc384d617e2cdf2186246988edc3130d5b: Status 404 returned error can't find the container with id 36620aaebd7df7c81208ba26284d4ffc384d617e2cdf2186246988edc3130d5b Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.192982 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" event={"ID":"8cb8d05a-e31a-4567-be65-b33c94f7c67f","Type":"ContainerStarted","Data":"61774122b39452d9a7b726ab592ee8780881d81a96e03a65e95bcbe212890942"} Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.193342 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" event={"ID":"8cb8d05a-e31a-4567-be65-b33c94f7c67f","Type":"ContainerStarted","Data":"36620aaebd7df7c81208ba26284d4ffc384d617e2cdf2186246988edc3130d5b"} Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.193809 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.195712 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" event={"ID":"d80beefa-ae96-4167-9bbe-6d2b0cfd8605","Type":"ContainerStarted","Data":"0e80428f16ab569957cf4f76391f915a4e6475ddec8def099a686c0d3e313efb"} Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.195761 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" event={"ID":"d80beefa-ae96-4167-9bbe-6d2b0cfd8605","Type":"ContainerStarted","Data":"b334087c67b446ad70991961647c2ee057cae06301a41b925cd8cf0e89126ad1"} Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.196621 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.200516 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.201655 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" Oct 01 07:18:15 crc kubenswrapper[4837]: I1001 07:18:15.215853 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bdd8c968c-h24zp" podStartSLOduration=4.215835353 podStartE2EDuration="4.215835353s" podCreationTimestamp="2025-10-01 07:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:18:15.213954778 +0000 UTC m=+752.055562283" watchObservedRunningTime="2025-10-01 07:18:15.215835353 +0000 UTC m=+752.057442818" Oct 01 07:18:18 crc kubenswrapper[4837]: I1001 07:18:18.670253 4837 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.079297 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.079630 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.079683 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.080285 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.080349 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05" gracePeriod=600 Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.252153 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05" exitCode=0 Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.252233 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05"} Oct 01 07:18:23 crc kubenswrapper[4837]: I1001 07:18:23.252632 4837 scope.go:117] "RemoveContainer" containerID="1ce67988ceb032b7615c1a5e51777585933d748b6e994296bd867e1717852a9b" Oct 01 07:18:24 crc kubenswrapper[4837]: I1001 07:18:24.261811 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb"} Oct 01 07:18:24 crc kubenswrapper[4837]: I1001 07:18:24.287276 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-675ff74c79-5cg2k" podStartSLOduration=13.287259044 podStartE2EDuration="13.287259044s" podCreationTimestamp="2025-10-01 07:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:18:15.277302445 +0000 UTC m=+752.118909940" watchObservedRunningTime="2025-10-01 07:18:24.287259044 +0000 UTC m=+761.128866509" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.019227 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vf2cm"] Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020462 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-controller" containerID="cri-o://8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020569 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="northd" containerID="cri-o://37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020624 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-acl-logging" containerID="cri-o://18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020617 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-node" containerID="cri-o://de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020666 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020801 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="nbdb" containerID="cri-o://6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.020849 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="sbdb" containerID="cri-o://a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.094730 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" containerID="cri-o://5b0c10fc9ea9b7cbb60afef0c27d6ccfed8c99619eb8bc8ba7a28d700b7e31b9" gracePeriod=30 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.294760 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovnkube-controller/3.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297038 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-acl-logging/0.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297506 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-controller/0.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297867 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="5b0c10fc9ea9b7cbb60afef0c27d6ccfed8c99619eb8bc8ba7a28d700b7e31b9" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297890 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297901 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297911 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297921 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297930 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f" exitCode=0 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297941 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c" exitCode=143 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297950 4837 generic.go:334] "Generic (PLEG): container finished" podID="401d9e9e-cba5-413c-b078-83858883db16" containerID="8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31" exitCode=143 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.297996 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"5b0c10fc9ea9b7cbb60afef0c27d6ccfed8c99619eb8bc8ba7a28d700b7e31b9"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298052 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298063 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298076 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298087 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298098 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298110 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" event={"ID":"401d9e9e-cba5-413c-b078-83858883db16","Type":"ContainerDied","Data":"99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298121 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99ece544435fba13e0b67519ebb5bd6db6facfa2db1eaa5f8513b6fa3a1301ee" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.298138 4837 scope.go:117] "RemoveContainer" containerID="f8a3b93d07a206fc95abf6177c9a9ede0f786d75014d6c156ffb89b625490e6c" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.301141 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/2.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.301910 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/1.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.301939 4837 generic.go:334] "Generic (PLEG): container finished" podID="7dedbee2-d2c0-49fb-ac76-aa7562c61211" containerID="efa528d87556ee38ed9f50eab56f31e7f63b2e759fdc058103598ba4f496886e" exitCode=2 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.301960 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerDied","Data":"efa528d87556ee38ed9f50eab56f31e7f63b2e759fdc058103598ba4f496886e"} Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.302357 4837 scope.go:117] "RemoveContainer" containerID="efa528d87556ee38ed9f50eab56f31e7f63b2e759fdc058103598ba4f496886e" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.347362 4837 scope.go:117] "RemoveContainer" containerID="f02a21365f1b7b01b36a599493b34785cecb1eafdc64b7506a0733d6e4497952" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.351634 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-acl-logging/0.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.352510 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-controller/0.log" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.353030 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403630 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hcjbn"] Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403865 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kubecfg-setup" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403878 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kubecfg-setup" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403893 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403900 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403912 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403919 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403927 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="northd" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403932 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="northd" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403939 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="sbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403944 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="sbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403956 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403962 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403971 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-node" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403976 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-node" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403984 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.403989 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.403998 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.404003 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.404010 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.404016 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.404022 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-acl-logging" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.404027 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-acl-logging" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.404034 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="nbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.404039 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="nbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405218 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-node" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405237 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405244 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405253 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405263 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="nbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405289 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405297 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405305 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405314 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="northd" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405321 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="sbdb" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405330 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovn-acl-logging" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.405472 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405482 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.405584 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="401d9e9e-cba5-413c-b078-83858883db16" containerName="ovnkube-controller" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.407337 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.468988 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469235 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469273 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469297 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469099 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469325 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469419 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469427 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket" (OuterVolumeSpecName: "log-socket") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469453 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469422 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469502 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469541 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469581 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469607 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469632 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5btcx\" (UniqueName: \"kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469636 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469659 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469696 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469719 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469754 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469774 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469801 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469816 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469844 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469873 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.469895 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch\") pod \"401d9e9e-cba5-413c-b078-83858883db16\" (UID: \"401d9e9e-cba5-413c-b078-83858883db16\") " Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470061 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-var-lib-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470095 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovn-node-metrics-cert\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470115 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-netns\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470137 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-slash\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470165 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-ovn\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470188 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-systemd-units\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470211 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-netd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470227 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jft96\" (UniqueName: \"kubernetes.io/projected/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-kube-api-access-jft96\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470264 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-bin\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470298 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-script-lib\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470354 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-log-socket\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470370 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-systemd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-config\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470420 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-env-overrides\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470444 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470502 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470547 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash" (OuterVolumeSpecName: "host-slash") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470549 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470575 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470582 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470579 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log" (OuterVolumeSpecName: "node-log") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470610 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470646 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470901 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470934 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-etc-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.470980 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-node-log\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471005 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-kubelet\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471027 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471088 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471131 4837 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471149 4837 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471158 4837 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471170 4837 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-log-socket\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471202 4837 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471210 4837 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471218 4837 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471226 4837 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471257 4837 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471267 4837 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-node-log\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471276 4837 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471283 4837 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471292 4837 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471300 4837 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471309 4837 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-slash\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.471316 4837 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.478190 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx" (OuterVolumeSpecName: "kube-api-access-5btcx") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "kube-api-access-5btcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.479145 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.487136 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "401d9e9e-cba5-413c-b078-83858883db16" (UID: "401d9e9e-cba5-413c-b078-83858883db16"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.571889 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-script-lib\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.571939 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.571967 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-log-socket\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.571991 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-systemd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-config\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572028 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-env-overrides\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572043 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572061 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-etc-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572078 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-node-log\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572094 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-kubelet\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572090 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-log-socket\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-systemd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572161 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572184 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-etc-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572158 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-kubelet\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-node-log\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572108 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572269 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-var-lib-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572333 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-var-lib-openvswitch\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572398 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovn-node-metrics-cert\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572422 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-netns\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572454 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-slash\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572492 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-ovn\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572538 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-run-netns\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572575 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-slash\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.572919 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-env-overrides\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573282 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-run-ovn\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573347 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-systemd-units\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573381 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jft96\" (UniqueName: \"kubernetes.io/projected/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-kube-api-access-jft96\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573400 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-netd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573412 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-systemd-units\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573419 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-bin\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573444 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-bin\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573527 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401d9e9e-cba5-413c-b078-83858883db16-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573544 4837 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401d9e9e-cba5-413c-b078-83858883db16-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573557 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5btcx\" (UniqueName: \"kubernetes.io/projected/401d9e9e-cba5-413c-b078-83858883db16-kube-api-access-5btcx\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573569 4837 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401d9e9e-cba5-413c-b078-83858883db16-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-host-cni-netd\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573618 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-script-lib\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.573741 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovnkube-config\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.576327 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-ovn-node-metrics-cert\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.604403 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.606892 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.611743 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jft96\" (UniqueName: \"kubernetes.io/projected/811d5074-eca2-4e2b-a8d6-1d133ed59c6d-kube-api-access-jft96\") pod \"ovnkube-node-hcjbn\" (UID: \"811d5074-eca2-4e2b-a8d6-1d133ed59c6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.675937 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.676126 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzzfn\" (UniqueName: \"kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.682835 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.729343 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:29 crc kubenswrapper[4837]: W1001 07:18:29.748184 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod811d5074_eca2_4e2b_a8d6_1d133ed59c6d.slice/crio-94473179788b1dc72816faf812454e5767d9225dd6b990d1a736dec3336e5c80 WatchSource:0}: Error finding container 94473179788b1dc72816faf812454e5767d9225dd6b990d1a736dec3336e5c80: Status 404 returned error can't find the container with id 94473179788b1dc72816faf812454e5767d9225dd6b990d1a736dec3336e5c80 Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.783946 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzzfn\" (UniqueName: \"kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.784009 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.784039 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.784458 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.784669 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.812310 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzzfn\" (UniqueName: \"kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn\") pod \"redhat-marketplace-z7k8p\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: I1001 07:18:29.923124 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.952646 4837 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(9c8138ad7ff2cfb0de5bb286f52ca5b2a61f8001ccf0c614aab2a83365fd1c46): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.952818 4837 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(9c8138ad7ff2cfb0de5bb286f52ca5b2a61f8001ccf0c614aab2a83365fd1c46): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.952918 4837 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(9c8138ad7ff2cfb0de5bb286f52ca5b2a61f8001ccf0c614aab2a83365fd1c46): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:29 crc kubenswrapper[4837]: E1001 07:18:29.953057 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-z7k8p_openshift-marketplace(4aa64e76-ef53-497f-9f7f-5f734adea7a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-z7k8p_openshift-marketplace(4aa64e76-ef53-497f-9f7f-5f734adea7a5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(9c8138ad7ff2cfb0de5bb286f52ca5b2a61f8001ccf0c614aab2a83365fd1c46): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-z7k8p" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.311657 4837 generic.go:334] "Generic (PLEG): container finished" podID="811d5074-eca2-4e2b-a8d6-1d133ed59c6d" containerID="feef994c80fd844e319fb6e5618dd5ea1fd5379f1ba083b791d1d95e63be4fa1" exitCode=0 Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.311776 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerDied","Data":"feef994c80fd844e319fb6e5618dd5ea1fd5379f1ba083b791d1d95e63be4fa1"} Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.311868 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"94473179788b1dc72816faf812454e5767d9225dd6b990d1a736dec3336e5c80"} Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.321402 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-acl-logging/0.log" Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.322221 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vf2cm_401d9e9e-cba5-413c-b078-83858883db16/ovn-controller/0.log" Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.322955 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vf2cm" Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.327890 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-kztv6_7dedbee2-d2c0-49fb-ac76-aa7562c61211/kube-multus/2.log" Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.328046 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-kztv6" event={"ID":"7dedbee2-d2c0-49fb-ac76-aa7562c61211","Type":"ContainerStarted","Data":"73f4c0f73454056a77877ee86f96b1848a392839d77f9451a6f5e88abf3839d7"} Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.390954 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vf2cm"] Oct 01 07:18:30 crc kubenswrapper[4837]: I1001 07:18:30.396934 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vf2cm"] Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336336 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"05b744ae5c0daef8e6e13b59bdfa0284882e6ac5169e42147f15b07d58d1845c"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336579 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"1c28811b5eeff830429f75d493696bd75c82fdc1b5ff53a6dc212b4a31918121"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336592 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"893fbffa680c3c187db6353be727f82ba56ff9311e25cd120788000c6b8c2fa8"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336602 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"4d11e431a13fd8f98b66d1fee280b31cf8517e5b7ac470ed2c5c6904ff4b39bd"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336612 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"1b9ae24a95da4efd68f3122eaed336432591a055022c032640cf49a8771c8329"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.336620 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"0942c78fff5fe2f1f75bc4f4249da0f18ff253600a456c7b484059812a96144c"} Oct 01 07:18:31 crc kubenswrapper[4837]: I1001 07:18:31.825187 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401d9e9e-cba5-413c-b078-83858883db16" path="/var/lib/kubelet/pods/401d9e9e-cba5-413c-b078-83858883db16/volumes" Oct 01 07:18:34 crc kubenswrapper[4837]: I1001 07:18:34.366659 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"e2c0a699c324195839543592837ffe808e0aea0e62781bba49304ff384aa0cef"} Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.393486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" event={"ID":"811d5074-eca2-4e2b-a8d6-1d133ed59c6d","Type":"ContainerStarted","Data":"5d6e93c34d310ee1a8145bb88b63ad597f7b1b41a777b4e137babaecc4abae09"} Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.394258 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.394619 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.425526 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.436251 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" podStartSLOduration=7.436233232 podStartE2EDuration="7.436233232s" podCreationTimestamp="2025-10-01 07:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:18:36.434202163 +0000 UTC m=+773.275809638" watchObservedRunningTime="2025-10-01 07:18:36.436233232 +0000 UTC m=+773.277840697" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.440767 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.496103 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.496351 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:36 crc kubenswrapper[4837]: I1001 07:18:36.497193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:36 crc kubenswrapper[4837]: E1001 07:18:36.548203 4837 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(7141b069f8289dda8b35567ff93fa644654ebd26f197eb93b2600437268a83b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:18:36 crc kubenswrapper[4837]: E1001 07:18:36.548269 4837 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(7141b069f8289dda8b35567ff93fa644654ebd26f197eb93b2600437268a83b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:36 crc kubenswrapper[4837]: E1001 07:18:36.548301 4837 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(7141b069f8289dda8b35567ff93fa644654ebd26f197eb93b2600437268a83b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:36 crc kubenswrapper[4837]: E1001 07:18:36.548351 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-z7k8p_openshift-marketplace(4aa64e76-ef53-497f-9f7f-5f734adea7a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-z7k8p_openshift-marketplace(4aa64e76-ef53-497f-9f7f-5f734adea7a5)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-z7k8p_openshift-marketplace_4aa64e76-ef53-497f-9f7f-5f734adea7a5_0(7141b069f8289dda8b35567ff93fa644654ebd26f197eb93b2600437268a83b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/redhat-marketplace-z7k8p" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" Oct 01 07:18:37 crc kubenswrapper[4837]: I1001 07:18:37.402208 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.307755 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-4fp78"] Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.309154 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.313809 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.313814 4837 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5xt95" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.314152 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.318780 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.327923 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4fp78"] Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.404823 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.404901 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.405205 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrjjl\" (UniqueName: \"kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.507007 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.507314 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.507607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrjjl\" (UniqueName: \"kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.507399 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.508668 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.543566 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrjjl\" (UniqueName: \"kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl\") pod \"crc-storage-crc-4fp78\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: I1001 07:18:39.642265 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: E1001 07:18:39.690204 4837 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4fp78_crc-storage_eff67284-dd9c-485e-bb92-38e108c5d2af_0(0713ab8fc568fd5a3a7d9f239df2cde3f8d19c0be98ec033c695a11c4f0e5684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 07:18:39 crc kubenswrapper[4837]: E1001 07:18:39.690322 4837 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4fp78_crc-storage_eff67284-dd9c-485e-bb92-38e108c5d2af_0(0713ab8fc568fd5a3a7d9f239df2cde3f8d19c0be98ec033c695a11c4f0e5684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: E1001 07:18:39.690371 4837 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4fp78_crc-storage_eff67284-dd9c-485e-bb92-38e108c5d2af_0(0713ab8fc568fd5a3a7d9f239df2cde3f8d19c0be98ec033c695a11c4f0e5684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:39 crc kubenswrapper[4837]: E1001 07:18:39.690478 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4fp78_crc-storage(eff67284-dd9c-485e-bb92-38e108c5d2af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4fp78_crc-storage(eff67284-dd9c-485e-bb92-38e108c5d2af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4fp78_crc-storage_eff67284-dd9c-485e-bb92-38e108c5d2af_0(0713ab8fc568fd5a3a7d9f239df2cde3f8d19c0be98ec033c695a11c4f0e5684): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4fp78" podUID="eff67284-dd9c-485e-bb92-38e108c5d2af" Oct 01 07:18:40 crc kubenswrapper[4837]: I1001 07:18:40.425552 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:40 crc kubenswrapper[4837]: I1001 07:18:40.426183 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:40 crc kubenswrapper[4837]: I1001 07:18:40.935379 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4fp78"] Oct 01 07:18:40 crc kubenswrapper[4837]: W1001 07:18:40.946159 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeff67284_dd9c_485e_bb92_38e108c5d2af.slice/crio-641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a WatchSource:0}: Error finding container 641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a: Status 404 returned error can't find the container with id 641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a Oct 01 07:18:40 crc kubenswrapper[4837]: I1001 07:18:40.950339 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.032719 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.034391 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.041622 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.129361 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.129427 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.129465 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2k5j\" (UniqueName: \"kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.230580 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.230656 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.230725 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2k5j\" (UniqueName: \"kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.231886 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.231944 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.260037 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2k5j\" (UniqueName: \"kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j\") pod \"certified-operators-cb8c5\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.353661 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.435944 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fp78" event={"ID":"eff67284-dd9c-485e-bb92-38e108c5d2af","Type":"ContainerStarted","Data":"641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a"} Oct 01 07:18:41 crc kubenswrapper[4837]: I1001 07:18:41.775580 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:42 crc kubenswrapper[4837]: W1001 07:18:42.159667 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc439bf3_1797_44df_b5a5_319713fe7f96.slice/crio-d9ba62dd53643acd0e736d4825d47923d6d8f92c6fdca1f397c72797a0ee3726 WatchSource:0}: Error finding container d9ba62dd53643acd0e736d4825d47923d6d8f92c6fdca1f397c72797a0ee3726: Status 404 returned error can't find the container with id d9ba62dd53643acd0e736d4825d47923d6d8f92c6fdca1f397c72797a0ee3726 Oct 01 07:18:42 crc kubenswrapper[4837]: I1001 07:18:42.445762 4837 generic.go:334] "Generic (PLEG): container finished" podID="eff67284-dd9c-485e-bb92-38e108c5d2af" containerID="03b8114fbe7ef55b0cc69022bd642af7495d794afe4e71e7f326201ca1e32c42" exitCode=0 Oct 01 07:18:42 crc kubenswrapper[4837]: I1001 07:18:42.445858 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fp78" event={"ID":"eff67284-dd9c-485e-bb92-38e108c5d2af","Type":"ContainerDied","Data":"03b8114fbe7ef55b0cc69022bd642af7495d794afe4e71e7f326201ca1e32c42"} Oct 01 07:18:42 crc kubenswrapper[4837]: I1001 07:18:42.451247 4837 generic.go:334] "Generic (PLEG): container finished" podID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerID="defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e" exitCode=0 Oct 01 07:18:42 crc kubenswrapper[4837]: I1001 07:18:42.451283 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerDied","Data":"defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e"} Oct 01 07:18:42 crc kubenswrapper[4837]: I1001 07:18:42.451300 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerStarted","Data":"d9ba62dd53643acd0e736d4825d47923d6d8f92c6fdca1f397c72797a0ee3726"} Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.458176 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerStarted","Data":"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360"} Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.820381 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.963497 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage\") pod \"eff67284-dd9c-485e-bb92-38e108c5d2af\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.963587 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt\") pod \"eff67284-dd9c-485e-bb92-38e108c5d2af\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.963766 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrjjl\" (UniqueName: \"kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl\") pod \"eff67284-dd9c-485e-bb92-38e108c5d2af\" (UID: \"eff67284-dd9c-485e-bb92-38e108c5d2af\") " Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.963794 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "eff67284-dd9c-485e-bb92-38e108c5d2af" (UID: "eff67284-dd9c-485e-bb92-38e108c5d2af"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.964052 4837 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/eff67284-dd9c-485e-bb92-38e108c5d2af-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.971566 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl" (OuterVolumeSpecName: "kube-api-access-nrjjl") pod "eff67284-dd9c-485e-bb92-38e108c5d2af" (UID: "eff67284-dd9c-485e-bb92-38e108c5d2af"). InnerVolumeSpecName "kube-api-access-nrjjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:43 crc kubenswrapper[4837]: I1001 07:18:43.988106 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "eff67284-dd9c-485e-bb92-38e108c5d2af" (UID: "eff67284-dd9c-485e-bb92-38e108c5d2af"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.065760 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrjjl\" (UniqueName: \"kubernetes.io/projected/eff67284-dd9c-485e-bb92-38e108c5d2af-kube-api-access-nrjjl\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.065803 4837 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/eff67284-dd9c-485e-bb92-38e108c5d2af-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.114004 4837 scope.go:117] "RemoveContainer" containerID="37ad971f29a23fa4f50ef75e3516564dfc43f40e460d7eb2006ae4e450ca0810" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.137105 4837 scope.go:117] "RemoveContainer" containerID="18c403dcb71357610a2009f5b2f91b8b179c51cb8fdfe1fc0e49a98ce9e3930c" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.159611 4837 scope.go:117] "RemoveContainer" containerID="cf2d695435740020cea074fe5651dd870f9f7b06eb22beacc07532707b6b2561" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.180963 4837 scope.go:117] "RemoveContainer" containerID="8b84b272ae1277b1bf2af1079b57e825e6df1ce019d77aa541971812c25f8e31" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.203457 4837 scope.go:117] "RemoveContainer" containerID="a30d1ddf3faff738ed27f0109266d1b391906f958db789a0d951ab49c1faedb8" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.223919 4837 scope.go:117] "RemoveContainer" containerID="5b0c10fc9ea9b7cbb60afef0c27d6ccfed8c99619eb8bc8ba7a28d700b7e31b9" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.242556 4837 scope.go:117] "RemoveContainer" containerID="6f353ff7e7ab535c045b27ce47f8cf1da520a041a60177fb76070212b41dfbf3" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.260105 4837 scope.go:117] "RemoveContainer" containerID="a6d2b109d7108682f5af4174c6901199bea67b61e2dc82c105a909e9efe9436a" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.276015 4837 scope.go:117] "RemoveContainer" containerID="de242629da3967ae9156642acfe405f0fecf788fbd354a70400f524461c90e2f" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.466111 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fp78" event={"ID":"eff67284-dd9c-485e-bb92-38e108c5d2af","Type":"ContainerDied","Data":"641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a"} Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.466854 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="641be7108d6876ea8b00500b5693f9fad54aca9437961c7c79f442fa45b8860a" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.466158 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fp78" Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.469118 4837 generic.go:334] "Generic (PLEG): container finished" podID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerID="8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360" exitCode=0 Oct 01 07:18:44 crc kubenswrapper[4837]: I1001 07:18:44.469201 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerDied","Data":"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360"} Oct 01 07:18:45 crc kubenswrapper[4837]: I1001 07:18:45.479441 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerStarted","Data":"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2"} Oct 01 07:18:45 crc kubenswrapper[4837]: I1001 07:18:45.504073 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cb8c5" podStartSLOduration=1.980616358 podStartE2EDuration="4.504049516s" podCreationTimestamp="2025-10-01 07:18:41 +0000 UTC" firstStartedPulling="2025-10-01 07:18:42.45280236 +0000 UTC m=+779.294409835" lastFinishedPulling="2025-10-01 07:18:44.976235498 +0000 UTC m=+781.817842993" observedRunningTime="2025-10-01 07:18:45.501900712 +0000 UTC m=+782.343508237" watchObservedRunningTime="2025-10-01 07:18:45.504049516 +0000 UTC m=+782.345657011" Oct 01 07:18:48 crc kubenswrapper[4837]: I1001 07:18:48.815755 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:48 crc kubenswrapper[4837]: I1001 07:18:48.816736 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:49 crc kubenswrapper[4837]: I1001 07:18:49.334923 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:18:49 crc kubenswrapper[4837]: W1001 07:18:49.343069 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aa64e76_ef53_497f_9f7f_5f734adea7a5.slice/crio-742e5177fc0b89b44bb02e64eebbbf232ffc5daa9cc59b097d4b688a2052de2e WatchSource:0}: Error finding container 742e5177fc0b89b44bb02e64eebbbf232ffc5daa9cc59b097d4b688a2052de2e: Status 404 returned error can't find the container with id 742e5177fc0b89b44bb02e64eebbbf232ffc5daa9cc59b097d4b688a2052de2e Oct 01 07:18:49 crc kubenswrapper[4837]: I1001 07:18:49.509340 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerStarted","Data":"742e5177fc0b89b44bb02e64eebbbf232ffc5daa9cc59b097d4b688a2052de2e"} Oct 01 07:18:50 crc kubenswrapper[4837]: I1001 07:18:50.540648 4837 generic.go:334] "Generic (PLEG): container finished" podID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerID="e294b1285c56f995d225ff5cba54e0fac468a62df78a08559947840db3e1019b" exitCode=0 Oct 01 07:18:50 crc kubenswrapper[4837]: I1001 07:18:50.540739 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerDied","Data":"e294b1285c56f995d225ff5cba54e0fac468a62df78a08559947840db3e1019b"} Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.354810 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.355194 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.396917 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.548489 4837 generic.go:334] "Generic (PLEG): container finished" podID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerID="6b5e68f1a81a6abf78519669fee69fedeab2c423e54cb3d105cbcf782758a4b1" exitCode=0 Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.548578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerDied","Data":"6b5e68f1a81a6abf78519669fee69fedeab2c423e54cb3d105cbcf782758a4b1"} Oct 01 07:18:51 crc kubenswrapper[4837]: I1001 07:18:51.591540 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.192683 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf"] Oct 01 07:18:52 crc kubenswrapper[4837]: E1001 07:18:52.193296 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff67284-dd9c-485e-bb92-38e108c5d2af" containerName="storage" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.193426 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff67284-dd9c-485e-bb92-38e108c5d2af" containerName="storage" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.193740 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff67284-dd9c-485e-bb92-38e108c5d2af" containerName="storage" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.194861 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.198068 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.213788 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf"] Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.274970 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.275025 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltsr4\" (UniqueName: \"kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.275071 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.376874 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.377181 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.377313 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltsr4\" (UniqueName: \"kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.377629 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.378016 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.413573 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltsr4\" (UniqueName: \"kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.512063 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.559115 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerStarted","Data":"860cf729c4e981d0f6af878f89c2cf7a3019ccb30f80c9524a9b5547780feef5"} Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.797380 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z7k8p" podStartSLOduration=22.354545978 podStartE2EDuration="23.797355348s" podCreationTimestamp="2025-10-01 07:18:29 +0000 UTC" firstStartedPulling="2025-10-01 07:18:50.545309575 +0000 UTC m=+787.386917030" lastFinishedPulling="2025-10-01 07:18:51.988118945 +0000 UTC m=+788.829726400" observedRunningTime="2025-10-01 07:18:52.587721825 +0000 UTC m=+789.429329280" watchObservedRunningTime="2025-10-01 07:18:52.797355348 +0000 UTC m=+789.638962813" Oct 01 07:18:52 crc kubenswrapper[4837]: I1001 07:18:52.801932 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf"] Oct 01 07:18:52 crc kubenswrapper[4837]: W1001 07:18:52.806736 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae36cbc_90d3_46bc_a06b_4dc3a17721c9.slice/crio-fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4 WatchSource:0}: Error finding container fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4: Status 404 returned error can't find the container with id fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4 Oct 01 07:18:53 crc kubenswrapper[4837]: I1001 07:18:53.569196 4837 generic.go:334] "Generic (PLEG): container finished" podID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerID="9b8a61143656840102e25cf698319bca535ffbe13c0f913b73f2f3d6612b2efd" exitCode=0 Oct 01 07:18:53 crc kubenswrapper[4837]: I1001 07:18:53.569343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" event={"ID":"bae36cbc-90d3-46bc-a06b-4dc3a17721c9","Type":"ContainerDied","Data":"9b8a61143656840102e25cf698319bca535ffbe13c0f913b73f2f3d6612b2efd"} Oct 01 07:18:53 crc kubenswrapper[4837]: I1001 07:18:53.569682 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" event={"ID":"bae36cbc-90d3-46bc-a06b-4dc3a17721c9","Type":"ContainerStarted","Data":"fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4"} Oct 01 07:18:53 crc kubenswrapper[4837]: I1001 07:18:53.724587 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:53 crc kubenswrapper[4837]: I1001 07:18:53.724955 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cb8c5" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="registry-server" containerID="cri-o://34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2" gracePeriod=2 Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.174970 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.328665 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities\") pod \"fc439bf3-1797-44df-b5a5-319713fe7f96\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.328854 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2k5j\" (UniqueName: \"kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j\") pod \"fc439bf3-1797-44df-b5a5-319713fe7f96\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.328928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content\") pod \"fc439bf3-1797-44df-b5a5-319713fe7f96\" (UID: \"fc439bf3-1797-44df-b5a5-319713fe7f96\") " Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.329859 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities" (OuterVolumeSpecName: "utilities") pod "fc439bf3-1797-44df-b5a5-319713fe7f96" (UID: "fc439bf3-1797-44df-b5a5-319713fe7f96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.340977 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j" (OuterVolumeSpecName: "kube-api-access-w2k5j") pod "fc439bf3-1797-44df-b5a5-319713fe7f96" (UID: "fc439bf3-1797-44df-b5a5-319713fe7f96"). InnerVolumeSpecName "kube-api-access-w2k5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.417619 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc439bf3-1797-44df-b5a5-319713fe7f96" (UID: "fc439bf3-1797-44df-b5a5-319713fe7f96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.430329 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.430362 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2k5j\" (UniqueName: \"kubernetes.io/projected/fc439bf3-1797-44df-b5a5-319713fe7f96-kube-api-access-w2k5j\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.430376 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc439bf3-1797-44df-b5a5-319713fe7f96-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.581067 4837 generic.go:334] "Generic (PLEG): container finished" podID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerID="34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2" exitCode=0 Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.581123 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerDied","Data":"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2"} Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.581162 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb8c5" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.581485 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb8c5" event={"ID":"fc439bf3-1797-44df-b5a5-319713fe7f96","Type":"ContainerDied","Data":"d9ba62dd53643acd0e736d4825d47923d6d8f92c6fdca1f397c72797a0ee3726"} Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.581544 4837 scope.go:117] "RemoveContainer" containerID="34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.609063 4837 scope.go:117] "RemoveContainer" containerID="8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.617378 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.620723 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cb8c5"] Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.700144 4837 scope.go:117] "RemoveContainer" containerID="defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.727030 4837 scope.go:117] "RemoveContainer" containerID="34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2" Oct 01 07:18:54 crc kubenswrapper[4837]: E1001 07:18:54.727551 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2\": container with ID starting with 34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2 not found: ID does not exist" containerID="34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.727584 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2"} err="failed to get container status \"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2\": rpc error: code = NotFound desc = could not find container \"34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2\": container with ID starting with 34f43daefee08f2c018aa97f6ae99d8177c7c7a73fdaa306da8a60151d2a4aa2 not found: ID does not exist" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.727609 4837 scope.go:117] "RemoveContainer" containerID="8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360" Oct 01 07:18:54 crc kubenswrapper[4837]: E1001 07:18:54.728213 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360\": container with ID starting with 8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360 not found: ID does not exist" containerID="8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.728236 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360"} err="failed to get container status \"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360\": rpc error: code = NotFound desc = could not find container \"8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360\": container with ID starting with 8d7281f13b5bb0182952ef64ed309bb47a41ea487685cb26e14d27b12a3c6360 not found: ID does not exist" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.728253 4837 scope.go:117] "RemoveContainer" containerID="defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e" Oct 01 07:18:54 crc kubenswrapper[4837]: E1001 07:18:54.729494 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e\": container with ID starting with defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e not found: ID does not exist" containerID="defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e" Oct 01 07:18:54 crc kubenswrapper[4837]: I1001 07:18:54.729518 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e"} err="failed to get container status \"defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e\": rpc error: code = NotFound desc = could not find container \"defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e\": container with ID starting with defd9547fc9c54085f52e46c94958853a867112a6eb5888d64c8505c399e001e not found: ID does not exist" Oct 01 07:18:55 crc kubenswrapper[4837]: I1001 07:18:55.591167 4837 generic.go:334] "Generic (PLEG): container finished" podID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerID="45041722717d2a993e41d6a330161e7a9fc1f5f192318e4cc813055081b3ff9b" exitCode=0 Oct 01 07:18:55 crc kubenswrapper[4837]: I1001 07:18:55.591273 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" event={"ID":"bae36cbc-90d3-46bc-a06b-4dc3a17721c9","Type":"ContainerDied","Data":"45041722717d2a993e41d6a330161e7a9fc1f5f192318e4cc813055081b3ff9b"} Oct 01 07:18:55 crc kubenswrapper[4837]: I1001 07:18:55.827153 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" path="/var/lib/kubelet/pods/fc439bf3-1797-44df-b5a5-319713fe7f96/volumes" Oct 01 07:18:56 crc kubenswrapper[4837]: I1001 07:18:56.604504 4837 generic.go:334] "Generic (PLEG): container finished" podID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerID="e58e39836b5c1d3475ee0997a636236713e05ac5e4fbedb6382180219b8f5b02" exitCode=0 Oct 01 07:18:56 crc kubenswrapper[4837]: I1001 07:18:56.604559 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" event={"ID":"bae36cbc-90d3-46bc-a06b-4dc3a17721c9","Type":"ContainerDied","Data":"e58e39836b5c1d3475ee0997a636236713e05ac5e4fbedb6382180219b8f5b02"} Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.339977 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:18:57 crc kubenswrapper[4837]: E1001 07:18:57.340333 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="registry-server" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.340355 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="registry-server" Oct 01 07:18:57 crc kubenswrapper[4837]: E1001 07:18:57.340375 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="extract-content" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.340387 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="extract-content" Oct 01 07:18:57 crc kubenswrapper[4837]: E1001 07:18:57.340405 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="extract-utilities" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.340420 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="extract-utilities" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.340642 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc439bf3-1797-44df-b5a5-319713fe7f96" containerName="registry-server" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.348567 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.350452 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.472475 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpn54\" (UniqueName: \"kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.472570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.472630 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.573738 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.573817 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.573888 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpn54\" (UniqueName: \"kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.574996 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.575732 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.602882 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpn54\" (UniqueName: \"kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54\") pod \"redhat-operators-hg9ht\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.672038 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.891332 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:18:57 crc kubenswrapper[4837]: W1001 07:18:57.895357 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e31c84e_1f47_412d_8557_14721386986f.slice/crio-299f487493f3122dbdd0efbc47f7e31f2b9eb79130929d972ccf088e5b4e1ef0 WatchSource:0}: Error finding container 299f487493f3122dbdd0efbc47f7e31f2b9eb79130929d972ccf088e5b4e1ef0: Status 404 returned error can't find the container with id 299f487493f3122dbdd0efbc47f7e31f2b9eb79130929d972ccf088e5b4e1ef0 Oct 01 07:18:57 crc kubenswrapper[4837]: I1001 07:18:57.903078 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.078984 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle\") pod \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.079206 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltsr4\" (UniqueName: \"kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4\") pod \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.079365 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util\") pod \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\" (UID: \"bae36cbc-90d3-46bc-a06b-4dc3a17721c9\") " Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.079803 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle" (OuterVolumeSpecName: "bundle") pod "bae36cbc-90d3-46bc-a06b-4dc3a17721c9" (UID: "bae36cbc-90d3-46bc-a06b-4dc3a17721c9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.084721 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4" (OuterVolumeSpecName: "kube-api-access-ltsr4") pod "bae36cbc-90d3-46bc-a06b-4dc3a17721c9" (UID: "bae36cbc-90d3-46bc-a06b-4dc3a17721c9"). InnerVolumeSpecName "kube-api-access-ltsr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.186247 4837 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.186328 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltsr4\" (UniqueName: \"kubernetes.io/projected/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-kube-api-access-ltsr4\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.376250 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util" (OuterVolumeSpecName: "util") pod "bae36cbc-90d3-46bc-a06b-4dc3a17721c9" (UID: "bae36cbc-90d3-46bc-a06b-4dc3a17721c9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.389382 4837 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bae36cbc-90d3-46bc-a06b-4dc3a17721c9-util\") on node \"crc\" DevicePath \"\"" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.617855 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e31c84e-1f47-412d-8557-14721386986f" containerID="09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e" exitCode=0 Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.617925 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerDied","Data":"09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e"} Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.617962 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerStarted","Data":"299f487493f3122dbdd0efbc47f7e31f2b9eb79130929d972ccf088e5b4e1ef0"} Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.621728 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" event={"ID":"bae36cbc-90d3-46bc-a06b-4dc3a17721c9","Type":"ContainerDied","Data":"fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4"} Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.621972 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb3e9de9df13835992eede59eaf8ccf8e3599f12eb28f6bfb2ef4ea346c037f4" Oct 01 07:18:58 crc kubenswrapper[4837]: I1001 07:18:58.622276 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf" Oct 01 07:18:59 crc kubenswrapper[4837]: I1001 07:18:59.766105 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcjbn" Oct 01 07:18:59 crc kubenswrapper[4837]: I1001 07:18:59.923493 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:59 crc kubenswrapper[4837]: I1001 07:18:59.923865 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:18:59 crc kubenswrapper[4837]: I1001 07:18:59.976642 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:19:00 crc kubenswrapper[4837]: I1001 07:19:00.642456 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e31c84e-1f47-412d-8557-14721386986f" containerID="762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5" exitCode=0 Oct 01 07:19:00 crc kubenswrapper[4837]: I1001 07:19:00.642543 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerDied","Data":"762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5"} Oct 01 07:19:00 crc kubenswrapper[4837]: I1001 07:19:00.714019 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:19:01 crc kubenswrapper[4837]: I1001 07:19:01.650138 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerStarted","Data":"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39"} Oct 01 07:19:01 crc kubenswrapper[4837]: I1001 07:19:01.668708 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hg9ht" podStartSLOduration=2.155011959 podStartE2EDuration="4.668672964s" podCreationTimestamp="2025-10-01 07:18:57 +0000 UTC" firstStartedPulling="2025-10-01 07:18:58.622883895 +0000 UTC m=+795.464491380" lastFinishedPulling="2025-10-01 07:19:01.13654489 +0000 UTC m=+797.978152385" observedRunningTime="2025-10-01 07:19:01.66688319 +0000 UTC m=+798.508490645" watchObservedRunningTime="2025-10-01 07:19:01.668672964 +0000 UTC m=+798.510280419" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.048787 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr"] Oct 01 07:19:02 crc kubenswrapper[4837]: E1001 07:19:02.049090 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="extract" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.049116 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="extract" Oct 01 07:19:02 crc kubenswrapper[4837]: E1001 07:19:02.049137 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="util" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.049168 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="util" Oct 01 07:19:02 crc kubenswrapper[4837]: E1001 07:19:02.049198 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="pull" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.049211 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="pull" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.049396 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae36cbc-90d3-46bc-a06b-4dc3a17721c9" containerName="extract" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.049967 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.055173 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.055429 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.056033 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-m8mqw" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.072009 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr"] Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.136553 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr8zg\" (UniqueName: \"kubernetes.io/projected/8898e3aa-b275-46de-92f9-1e1dd904d9bc-kube-api-access-tr8zg\") pod \"nmstate-operator-5d6f6cfd66-v5nzr\" (UID: \"8898e3aa-b275-46de-92f9-1e1dd904d9bc\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.237743 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr8zg\" (UniqueName: \"kubernetes.io/projected/8898e3aa-b275-46de-92f9-1e1dd904d9bc-kube-api-access-tr8zg\") pod \"nmstate-operator-5d6f6cfd66-v5nzr\" (UID: \"8898e3aa-b275-46de-92f9-1e1dd904d9bc\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.260735 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr8zg\" (UniqueName: \"kubernetes.io/projected/8898e3aa-b275-46de-92f9-1e1dd904d9bc-kube-api-access-tr8zg\") pod \"nmstate-operator-5d6f6cfd66-v5nzr\" (UID: \"8898e3aa-b275-46de-92f9-1e1dd904d9bc\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.371024 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.615819 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr"] Oct 01 07:19:02 crc kubenswrapper[4837]: I1001 07:19:02.656009 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" event={"ID":"8898e3aa-b275-46de-92f9-1e1dd904d9bc","Type":"ContainerStarted","Data":"ac543d6a8b450bfdc8ae580560287114a1fb150e3cb9c2722857fc6fd70f79a4"} Oct 01 07:19:04 crc kubenswrapper[4837]: I1001 07:19:04.122024 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:19:04 crc kubenswrapper[4837]: I1001 07:19:04.122266 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z7k8p" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="registry-server" containerID="cri-o://860cf729c4e981d0f6af878f89c2cf7a3019ccb30f80c9524a9b5547780feef5" gracePeriod=2 Oct 01 07:19:04 crc kubenswrapper[4837]: I1001 07:19:04.681220 4837 generic.go:334] "Generic (PLEG): container finished" podID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerID="860cf729c4e981d0f6af878f89c2cf7a3019ccb30f80c9524a9b5547780feef5" exitCode=0 Oct 01 07:19:04 crc kubenswrapper[4837]: I1001 07:19:04.681503 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerDied","Data":"860cf729c4e981d0f6af878f89c2cf7a3019ccb30f80c9524a9b5547780feef5"} Oct 01 07:19:04 crc kubenswrapper[4837]: I1001 07:19:04.900637 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.078000 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content\") pod \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.078065 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities\") pod \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.078124 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzzfn\" (UniqueName: \"kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn\") pod \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\" (UID: \"4aa64e76-ef53-497f-9f7f-5f734adea7a5\") " Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.079007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities" (OuterVolumeSpecName: "utilities") pod "4aa64e76-ef53-497f-9f7f-5f734adea7a5" (UID: "4aa64e76-ef53-497f-9f7f-5f734adea7a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.083078 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn" (OuterVolumeSpecName: "kube-api-access-dzzfn") pod "4aa64e76-ef53-497f-9f7f-5f734adea7a5" (UID: "4aa64e76-ef53-497f-9f7f-5f734adea7a5"). InnerVolumeSpecName "kube-api-access-dzzfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.091575 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4aa64e76-ef53-497f-9f7f-5f734adea7a5" (UID: "4aa64e76-ef53-497f-9f7f-5f734adea7a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.179997 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.180053 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa64e76-ef53-497f-9f7f-5f734adea7a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.180073 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzzfn\" (UniqueName: \"kubernetes.io/projected/4aa64e76-ef53-497f-9f7f-5f734adea7a5-kube-api-access-dzzfn\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.689286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" event={"ID":"8898e3aa-b275-46de-92f9-1e1dd904d9bc","Type":"ContainerStarted","Data":"21b081d5cb7775f34fd6116c910b42d8bfb5b39b69bb4bfd2db011a2adfb76f8"} Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.692179 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7k8p" event={"ID":"4aa64e76-ef53-497f-9f7f-5f734adea7a5","Type":"ContainerDied","Data":"742e5177fc0b89b44bb02e64eebbbf232ffc5daa9cc59b097d4b688a2052de2e"} Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.692246 4837 scope.go:117] "RemoveContainer" containerID="860cf729c4e981d0f6af878f89c2cf7a3019ccb30f80c9524a9b5547780feef5" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.692312 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7k8p" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.711636 4837 scope.go:117] "RemoveContainer" containerID="6b5e68f1a81a6abf78519669fee69fedeab2c423e54cb3d105cbcf782758a4b1" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.712232 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-v5nzr" podStartSLOduration=1.698683094 podStartE2EDuration="3.712220033s" podCreationTimestamp="2025-10-01 07:19:02 +0000 UTC" firstStartedPulling="2025-10-01 07:19:02.63862837 +0000 UTC m=+799.480235825" lastFinishedPulling="2025-10-01 07:19:04.652165309 +0000 UTC m=+801.493772764" observedRunningTime="2025-10-01 07:19:05.711091965 +0000 UTC m=+802.552699420" watchObservedRunningTime="2025-10-01 07:19:05.712220033 +0000 UTC m=+802.553827478" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.745585 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.752026 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7k8p"] Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.754584 4837 scope.go:117] "RemoveContainer" containerID="e294b1285c56f995d225ff5cba54e0fac468a62df78a08559947840db3e1019b" Oct 01 07:19:05 crc kubenswrapper[4837]: I1001 07:19:05.821784 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" path="/var/lib/kubelet/pods/4aa64e76-ef53-497f-9f7f-5f734adea7a5/volumes" Oct 01 07:19:07 crc kubenswrapper[4837]: I1001 07:19:07.673027 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:07 crc kubenswrapper[4837]: I1001 07:19:07.673090 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:07 crc kubenswrapper[4837]: I1001 07:19:07.743928 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:07 crc kubenswrapper[4837]: I1001 07:19:07.825388 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:10 crc kubenswrapper[4837]: I1001 07:19:10.924617 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:19:10 crc kubenswrapper[4837]: I1001 07:19:10.925339 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hg9ht" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="registry-server" containerID="cri-o://ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39" gracePeriod=2 Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.354096 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.459968 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities\") pod \"3e31c84e-1f47-412d-8557-14721386986f\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.460127 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpn54\" (UniqueName: \"kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54\") pod \"3e31c84e-1f47-412d-8557-14721386986f\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.460207 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content\") pod \"3e31c84e-1f47-412d-8557-14721386986f\" (UID: \"3e31c84e-1f47-412d-8557-14721386986f\") " Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.461874 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities" (OuterVolumeSpecName: "utilities") pod "3e31c84e-1f47-412d-8557-14721386986f" (UID: "3e31c84e-1f47-412d-8557-14721386986f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.468987 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54" (OuterVolumeSpecName: "kube-api-access-bpn54") pod "3e31c84e-1f47-412d-8557-14721386986f" (UID: "3e31c84e-1f47-412d-8557-14721386986f"). InnerVolumeSpecName "kube-api-access-bpn54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.547357 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e31c84e-1f47-412d-8557-14721386986f" (UID: "3e31c84e-1f47-412d-8557-14721386986f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.561575 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpn54\" (UniqueName: \"kubernetes.io/projected/3e31c84e-1f47-412d-8557-14721386986f-kube-api-access-bpn54\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.561627 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.561646 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e31c84e-1f47-412d-8557-14721386986f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.736982 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e31c84e-1f47-412d-8557-14721386986f" containerID="ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39" exitCode=0 Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.737051 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerDied","Data":"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39"} Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.737114 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hg9ht" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.737150 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hg9ht" event={"ID":"3e31c84e-1f47-412d-8557-14721386986f","Type":"ContainerDied","Data":"299f487493f3122dbdd0efbc47f7e31f2b9eb79130929d972ccf088e5b4e1ef0"} Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.737206 4837 scope.go:117] "RemoveContainer" containerID="ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.770469 4837 scope.go:117] "RemoveContainer" containerID="762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.799167 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.804615 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hg9ht"] Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.807903 4837 scope.go:117] "RemoveContainer" containerID="09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.823930 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e31c84e-1f47-412d-8557-14721386986f" path="/var/lib/kubelet/pods/3e31c84e-1f47-412d-8557-14721386986f/volumes" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.824147 4837 scope.go:117] "RemoveContainer" containerID="ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39" Oct 01 07:19:11 crc kubenswrapper[4837]: E1001 07:19:11.824433 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39\": container with ID starting with ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39 not found: ID does not exist" containerID="ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.824460 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39"} err="failed to get container status \"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39\": rpc error: code = NotFound desc = could not find container \"ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39\": container with ID starting with ef27ecef95448b087c60cd6dcb17b4997327185de61935bccbfe077576b26a39 not found: ID does not exist" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.824483 4837 scope.go:117] "RemoveContainer" containerID="762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5" Oct 01 07:19:11 crc kubenswrapper[4837]: E1001 07:19:11.824745 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5\": container with ID starting with 762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5 not found: ID does not exist" containerID="762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.824777 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5"} err="failed to get container status \"762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5\": rpc error: code = NotFound desc = could not find container \"762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5\": container with ID starting with 762d396ad8fefd4e0fc71d384c6a9ad1cc5dcaf5ce72419638c5b34552774cf5 not found: ID does not exist" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.824795 4837 scope.go:117] "RemoveContainer" containerID="09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e" Oct 01 07:19:11 crc kubenswrapper[4837]: E1001 07:19:11.825020 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e\": container with ID starting with 09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e not found: ID does not exist" containerID="09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e" Oct 01 07:19:11 crc kubenswrapper[4837]: I1001 07:19:11.825058 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e"} err="failed to get container status \"09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e\": rpc error: code = NotFound desc = could not find container \"09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e\": container with ID starting with 09e9bc6fa36175ff0d50802b24bb2ec8623f60aec9dc131d835781bdce8d8a9e not found: ID does not exist" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.167624 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-47mbv"] Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.167911 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="extract-utilities" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.167931 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="extract-utilities" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.167952 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="extract-content" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.167962 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="extract-content" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.167976 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.167987 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.168006 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="extract-utilities" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.168015 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="extract-utilities" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.168030 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.168039 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.168054 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="extract-content" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.168063 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="extract-content" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.168212 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa64e76-ef53-497f-9f7f-5f734adea7a5" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.168237 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e31c84e-1f47-412d-8557-14721386986f" containerName="registry-server" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.169108 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.169979 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crbrz\" (UniqueName: \"kubernetes.io/projected/b62aa771-d06e-4b59-ab29-ad40a92198ee-kube-api-access-crbrz\") pod \"nmstate-metrics-58fcddf996-47mbv\" (UID: \"b62aa771-d06e-4b59-ab29-ad40a92198ee\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.178642 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fcvpd" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.183047 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-bcn95"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.183851 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.186920 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.196113 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-47mbv"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.208680 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-bcn95"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.271172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crbrz\" (UniqueName: \"kubernetes.io/projected/b62aa771-d06e-4b59-ab29-ad40a92198ee-kube-api-access-crbrz\") pod \"nmstate-metrics-58fcddf996-47mbv\" (UID: \"b62aa771-d06e-4b59-ab29-ad40a92198ee\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.281151 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-pwg46"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.281972 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.322726 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crbrz\" (UniqueName: \"kubernetes.io/projected/b62aa771-d06e-4b59-ab29-ad40a92198ee-kube-api-access-crbrz\") pod \"nmstate-metrics-58fcddf996-47mbv\" (UID: \"b62aa771-d06e-4b59-ab29-ad40a92198ee\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.352427 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.353086 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.359897 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.359898 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.359975 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-rpjt6" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.362368 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372370 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-nmstate-lock\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372429 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372465 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvsl4\" (UniqueName: \"kubernetes.io/projected/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-kube-api-access-jvsl4\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-ovs-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372566 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-dbus-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372618 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vfb\" (UniqueName: \"kubernetes.io/projected/71c13aee-6f73-4172-9a40-805d4b0c45bc-kube-api-access-f5vfb\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372677 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mtlw\" (UniqueName: \"kubernetes.io/projected/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-kube-api-access-4mtlw\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.372754 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473675 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mtlw\" (UniqueName: \"kubernetes.io/projected/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-kube-api-access-4mtlw\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473752 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473792 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-nmstate-lock\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473824 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473860 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvsl4\" (UniqueName: \"kubernetes.io/projected/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-kube-api-access-jvsl4\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473882 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-ovs-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473923 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473944 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-dbus-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.473970 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vfb\" (UniqueName: \"kubernetes.io/projected/71c13aee-6f73-4172-9a40-805d4b0c45bc-kube-api-access-f5vfb\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.474317 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-ovs-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.474392 4837 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.474437 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair podName:71c13aee-6f73-4172-9a40-805d4b0c45bc nodeName:}" failed. No retries permitted until 2025-10-01 07:19:12.974420196 +0000 UTC m=+809.816027651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair") pod "nmstate-webhook-6d689559c5-bcn95" (UID: "71c13aee-6f73-4172-9a40-805d4b0c45bc") : secret "openshift-nmstate-webhook" not found Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.474443 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-nmstate-lock\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.474564 4837 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 01 07:19:12 crc kubenswrapper[4837]: E1001 07:19:12.474649 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert podName:f00c5dd2-4cb3-4934-86d1-7fcd021441e1 nodeName:}" failed. No retries permitted until 2025-10-01 07:19:12.97460492 +0000 UTC m=+809.816212405 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-lbjd2" (UID: "f00c5dd2-4cb3-4934-86d1-7fcd021441e1") : secret "plugin-serving-cert" not found Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.474748 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.474866 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-dbus-socket\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.484821 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.499312 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mtlw\" (UniqueName: \"kubernetes.io/projected/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-kube-api-access-4mtlw\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.500374 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvsl4\" (UniqueName: \"kubernetes.io/projected/2e35d01a-91a7-4d3b-b9de-c688b3a7f10e-kube-api-access-jvsl4\") pod \"nmstate-handler-pwg46\" (UID: \"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e\") " pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.507339 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vfb\" (UniqueName: \"kubernetes.io/projected/71c13aee-6f73-4172-9a40-805d4b0c45bc-kube-api-access-f5vfb\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.570727 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-648c97fcb7-jf6rh"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.572269 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.580546 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-oauth-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.580604 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-service-ca\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.580654 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.580681 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4wh\" (UniqueName: \"kubernetes.io/projected/a7b8b81e-4029-451c-8ae3-b83251af96ce-kube-api-access-9c4wh\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.580727 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-oauth-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.581013 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-trusted-ca-bundle\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.581060 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.593107 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-648c97fcb7-jf6rh"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.596387 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682650 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-oauth-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-service-ca\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682711 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682731 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4wh\" (UniqueName: \"kubernetes.io/projected/a7b8b81e-4029-451c-8ae3-b83251af96ce-kube-api-access-9c4wh\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682750 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-oauth-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.682794 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-trusted-ca-bundle\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.683588 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-trusted-ca-bundle\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.683726 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.684308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-service-ca\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.686979 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a7b8b81e-4029-451c-8ae3-b83251af96ce-oauth-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.697019 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-oauth-config\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.697460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8b81e-4029-451c-8ae3-b83251af96ce-console-serving-cert\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.703488 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4wh\" (UniqueName: \"kubernetes.io/projected/a7b8b81e-4029-451c-8ae3-b83251af96ce-kube-api-access-9c4wh\") pod \"console-648c97fcb7-jf6rh\" (UID: \"a7b8b81e-4029-451c-8ae3-b83251af96ce\") " pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.745115 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-47mbv"] Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.746318 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-pwg46" event={"ID":"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e","Type":"ContainerStarted","Data":"81af90304d9ed796d95cf9f8dc8fd2c2adecafb8a18902f67b775137a5d26aa7"} Oct 01 07:19:12 crc kubenswrapper[4837]: W1001 07:19:12.755150 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb62aa771_d06e_4b59_ab29_ad40a92198ee.slice/crio-5742c773001dddf36193844234d02b88732c31b3d658ea100027d56764aa40c8 WatchSource:0}: Error finding container 5742c773001dddf36193844234d02b88732c31b3d658ea100027d56764aa40c8: Status 404 returned error can't find the container with id 5742c773001dddf36193844234d02b88732c31b3d658ea100027d56764aa40c8 Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.917255 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.988126 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.988272 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.994719 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/71c13aee-6f73-4172-9a40-805d4b0c45bc-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-bcn95\" (UID: \"71c13aee-6f73-4172-9a40-805d4b0c45bc\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:12 crc kubenswrapper[4837]: I1001 07:19:12.995777 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f00c5dd2-4cb3-4934-86d1-7fcd021441e1-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-lbjd2\" (UID: \"f00c5dd2-4cb3-4934-86d1-7fcd021441e1\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.099270 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.270796 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.444475 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-648c97fcb7-jf6rh"] Oct 01 07:19:13 crc kubenswrapper[4837]: W1001 07:19:13.446822 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7b8b81e_4029_451c_8ae3_b83251af96ce.slice/crio-24edcc1b07631f7a66fa0e47dde1fda80f36942a899a692cfd395bda1380a36f WatchSource:0}: Error finding container 24edcc1b07631f7a66fa0e47dde1fda80f36942a899a692cfd395bda1380a36f: Status 404 returned error can't find the container with id 24edcc1b07631f7a66fa0e47dde1fda80f36942a899a692cfd395bda1380a36f Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.521852 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-bcn95"] Oct 01 07:19:13 crc kubenswrapper[4837]: W1001 07:19:13.534099 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71c13aee_6f73_4172_9a40_805d4b0c45bc.slice/crio-f2573ff85ee06f4c5d9c649a514f62bcaeedcd7bb819a911a498fdfc36a22f8a WatchSource:0}: Error finding container f2573ff85ee06f4c5d9c649a514f62bcaeedcd7bb819a911a498fdfc36a22f8a: Status 404 returned error can't find the container with id f2573ff85ee06f4c5d9c649a514f62bcaeedcd7bb819a911a498fdfc36a22f8a Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.537398 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2"] Oct 01 07:19:13 crc kubenswrapper[4837]: W1001 07:19:13.559908 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf00c5dd2_4cb3_4934_86d1_7fcd021441e1.slice/crio-169a34eddfb701c5d326db90df94ac2a23ab35f1cc8352de64e48a904fe7e600 WatchSource:0}: Error finding container 169a34eddfb701c5d326db90df94ac2a23ab35f1cc8352de64e48a904fe7e600: Status 404 returned error can't find the container with id 169a34eddfb701c5d326db90df94ac2a23ab35f1cc8352de64e48a904fe7e600 Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.762095 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" event={"ID":"f00c5dd2-4cb3-4934-86d1-7fcd021441e1","Type":"ContainerStarted","Data":"169a34eddfb701c5d326db90df94ac2a23ab35f1cc8352de64e48a904fe7e600"} Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.765312 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-648c97fcb7-jf6rh" event={"ID":"a7b8b81e-4029-451c-8ae3-b83251af96ce","Type":"ContainerStarted","Data":"3e4a1457b18489ceb7a8ec344b2e83cae68d00aeb82600655946b037b262cda6"} Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.765394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-648c97fcb7-jf6rh" event={"ID":"a7b8b81e-4029-451c-8ae3-b83251af96ce","Type":"ContainerStarted","Data":"24edcc1b07631f7a66fa0e47dde1fda80f36942a899a692cfd395bda1380a36f"} Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.770093 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" event={"ID":"b62aa771-d06e-4b59-ab29-ad40a92198ee","Type":"ContainerStarted","Data":"5742c773001dddf36193844234d02b88732c31b3d658ea100027d56764aa40c8"} Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.771888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" event={"ID":"71c13aee-6f73-4172-9a40-805d4b0c45bc","Type":"ContainerStarted","Data":"f2573ff85ee06f4c5d9c649a514f62bcaeedcd7bb819a911a498fdfc36a22f8a"} Oct 01 07:19:13 crc kubenswrapper[4837]: I1001 07:19:13.792340 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-648c97fcb7-jf6rh" podStartSLOduration=1.792308907 podStartE2EDuration="1.792308907s" podCreationTimestamp="2025-10-01 07:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:19:13.789416665 +0000 UTC m=+810.631024160" watchObservedRunningTime="2025-10-01 07:19:13.792308907 +0000 UTC m=+810.633916382" Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.792830 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" event={"ID":"71c13aee-6f73-4172-9a40-805d4b0c45bc","Type":"ContainerStarted","Data":"31c5d371a9ba0776f585f4ed5255a3fe7dcf74d426aa80bf047c32d67dd09903"} Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.793737 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.794499 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" event={"ID":"b62aa771-d06e-4b59-ab29-ad40a92198ee","Type":"ContainerStarted","Data":"92aa5ab9e94c695c8955a7511eb792f23899b8db391f8268a3f1820bec26b16d"} Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.797160 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-pwg46" event={"ID":"2e35d01a-91a7-4d3b-b9de-c688b3a7f10e","Type":"ContainerStarted","Data":"533ce558c8c2d273714697269cf4cf1f13d3cad7a8a8a6e38c999d7f572d5b23"} Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.797452 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.809007 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" podStartSLOduration=2.170469149 podStartE2EDuration="3.808991335s" podCreationTimestamp="2025-10-01 07:19:12 +0000 UTC" firstStartedPulling="2025-10-01 07:19:13.537080536 +0000 UTC m=+810.378687991" lastFinishedPulling="2025-10-01 07:19:15.175602682 +0000 UTC m=+812.017210177" observedRunningTime="2025-10-01 07:19:15.808112813 +0000 UTC m=+812.649720268" watchObservedRunningTime="2025-10-01 07:19:15.808991335 +0000 UTC m=+812.650598790" Oct 01 07:19:15 crc kubenswrapper[4837]: I1001 07:19:15.826704 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-pwg46" podStartSLOduration=1.277565286 podStartE2EDuration="3.826672696s" podCreationTimestamp="2025-10-01 07:19:12 +0000 UTC" firstStartedPulling="2025-10-01 07:19:12.626229695 +0000 UTC m=+809.467837140" lastFinishedPulling="2025-10-01 07:19:15.175337065 +0000 UTC m=+812.016944550" observedRunningTime="2025-10-01 07:19:15.826295037 +0000 UTC m=+812.667902492" watchObservedRunningTime="2025-10-01 07:19:15.826672696 +0000 UTC m=+812.668280151" Oct 01 07:19:16 crc kubenswrapper[4837]: I1001 07:19:16.804449 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" event={"ID":"f00c5dd2-4cb3-4934-86d1-7fcd021441e1","Type":"ContainerStarted","Data":"dd6a3651f93dd8335f364a00e612024c737f22e6bd82839cf45546d3d2d53c93"} Oct 01 07:19:16 crc kubenswrapper[4837]: I1001 07:19:16.830375 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-lbjd2" podStartSLOduration=2.242506337 podStartE2EDuration="4.830320383s" podCreationTimestamp="2025-10-01 07:19:12 +0000 UTC" firstStartedPulling="2025-10-01 07:19:13.564735106 +0000 UTC m=+810.406342551" lastFinishedPulling="2025-10-01 07:19:16.152549142 +0000 UTC m=+812.994156597" observedRunningTime="2025-10-01 07:19:16.826742013 +0000 UTC m=+813.668349468" watchObservedRunningTime="2025-10-01 07:19:16.830320383 +0000 UTC m=+813.671927848" Oct 01 07:19:17 crc kubenswrapper[4837]: I1001 07:19:17.813840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" event={"ID":"b62aa771-d06e-4b59-ab29-ad40a92198ee","Type":"ContainerStarted","Data":"1c4c45db9ac36004f8d022732a6b07004545326c35aecbf20d03f91f786e870d"} Oct 01 07:19:17 crc kubenswrapper[4837]: I1001 07:19:17.867827 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-47mbv" podStartSLOduration=1.11143945 podStartE2EDuration="5.867792914s" podCreationTimestamp="2025-10-01 07:19:12 +0000 UTC" firstStartedPulling="2025-10-01 07:19:12.758298133 +0000 UTC m=+809.599905588" lastFinishedPulling="2025-10-01 07:19:17.514651597 +0000 UTC m=+814.356259052" observedRunningTime="2025-10-01 07:19:17.865609979 +0000 UTC m=+814.707217494" watchObservedRunningTime="2025-10-01 07:19:17.867792914 +0000 UTC m=+814.709400429" Oct 01 07:19:22 crc kubenswrapper[4837]: I1001 07:19:22.631304 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-pwg46" Oct 01 07:19:22 crc kubenswrapper[4837]: I1001 07:19:22.918534 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:22 crc kubenswrapper[4837]: I1001 07:19:22.919094 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:22 crc kubenswrapper[4837]: I1001 07:19:22.927131 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:23 crc kubenswrapper[4837]: I1001 07:19:23.869034 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-648c97fcb7-jf6rh" Oct 01 07:19:23 crc kubenswrapper[4837]: I1001 07:19:23.949326 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:19:33 crc kubenswrapper[4837]: I1001 07:19:33.108530 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-bcn95" Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.750319 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.754800 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.769915 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.922067 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.922155 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:35 crc kubenswrapper[4837]: I1001 07:19:35.922309 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv4bz\" (UniqueName: \"kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.023170 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.023227 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.023253 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv4bz\" (UniqueName: \"kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.023951 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.024164 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.065933 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv4bz\" (UniqueName: \"kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz\") pod \"community-operators-qzf77\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.078324 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.535508 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.954621 4837 generic.go:334] "Generic (PLEG): container finished" podID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerID="69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921" exitCode=0 Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.955203 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerDied","Data":"69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921"} Oct 01 07:19:36 crc kubenswrapper[4837]: I1001 07:19:36.955262 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerStarted","Data":"d72aac1e62c282a4b356ce484bfcebc928992d5430312979b08b4a24bfb71327"} Oct 01 07:19:37 crc kubenswrapper[4837]: I1001 07:19:37.979024 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerStarted","Data":"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e"} Oct 01 07:19:38 crc kubenswrapper[4837]: I1001 07:19:38.988890 4837 generic.go:334] "Generic (PLEG): container finished" podID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerID="efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e" exitCode=0 Oct 01 07:19:38 crc kubenswrapper[4837]: I1001 07:19:38.988978 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerDied","Data":"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e"} Oct 01 07:19:39 crc kubenswrapper[4837]: I1001 07:19:39.997793 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerStarted","Data":"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30"} Oct 01 07:19:40 crc kubenswrapper[4837]: I1001 07:19:40.022084 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qzf77" podStartSLOduration=2.500780142 podStartE2EDuration="5.022053216s" podCreationTimestamp="2025-10-01 07:19:35 +0000 UTC" firstStartedPulling="2025-10-01 07:19:36.956961925 +0000 UTC m=+833.798569420" lastFinishedPulling="2025-10-01 07:19:39.478235039 +0000 UTC m=+836.319842494" observedRunningTime="2025-10-01 07:19:40.01980911 +0000 UTC m=+836.861416645" watchObservedRunningTime="2025-10-01 07:19:40.022053216 +0000 UTC m=+836.863660701" Oct 01 07:19:46 crc kubenswrapper[4837]: I1001 07:19:46.079104 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:46 crc kubenswrapper[4837]: I1001 07:19:46.079840 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:46 crc kubenswrapper[4837]: I1001 07:19:46.158339 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:47 crc kubenswrapper[4837]: I1001 07:19:47.128172 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:47 crc kubenswrapper[4837]: I1001 07:19:47.198555 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.014034 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mg4h6" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" containerID="cri-o://1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc" gracePeriod=15 Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.063635 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qzf77" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="registry-server" containerID="cri-o://521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30" gracePeriod=2 Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.465976 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mg4h6_c5794771-53a6-46dd-bd10-b37f6534b87b/console/0.log" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.466368 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.475750 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.609973 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610057 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610140 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610201 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities\") pod \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610235 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610306 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610344 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content\") pod \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610420 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610462 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znr77\" (UniqueName: \"kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77\") pod \"c5794771-53a6-46dd-bd10-b37f6534b87b\" (UID: \"c5794771-53a6-46dd-bd10-b37f6534b87b\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.610531 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv4bz\" (UniqueName: \"kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz\") pod \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\" (UID: \"a471a5ad-fe42-42a1-b2d6-3958fad91f5c\") " Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.611063 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities" (OuterVolumeSpecName: "utilities") pod "a471a5ad-fe42-42a1-b2d6-3958fad91f5c" (UID: "a471a5ad-fe42-42a1-b2d6-3958fad91f5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.611577 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config" (OuterVolumeSpecName: "console-config") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.616362 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.616605 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.617156 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca" (OuterVolumeSpecName: "service-ca") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.620646 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.620955 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77" (OuterVolumeSpecName: "kube-api-access-znr77") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "kube-api-access-znr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.621170 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c5794771-53a6-46dd-bd10-b37f6534b87b" (UID: "c5794771-53a6-46dd-bd10-b37f6534b87b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.621487 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz" (OuterVolumeSpecName: "kube-api-access-bv4bz") pod "a471a5ad-fe42-42a1-b2d6-3958fad91f5c" (UID: "a471a5ad-fe42-42a1-b2d6-3958fad91f5c"). InnerVolumeSpecName "kube-api-access-bv4bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.679538 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a471a5ad-fe42-42a1-b2d6-3958fad91f5c" (UID: "a471a5ad-fe42-42a1-b2d6-3958fad91f5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713195 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv4bz\" (UniqueName: \"kubernetes.io/projected/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-kube-api-access-bv4bz\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713243 4837 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713258 4837 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713272 4837 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713287 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713300 4837 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713313 4837 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5794771-53a6-46dd-bd10-b37f6534b87b-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713328 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a471a5ad-fe42-42a1-b2d6-3958fad91f5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713341 4837 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5794771-53a6-46dd-bd10-b37f6534b87b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:49 crc kubenswrapper[4837]: I1001 07:19:49.713353 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znr77\" (UniqueName: \"kubernetes.io/projected/c5794771-53a6-46dd-bd10-b37f6534b87b-kube-api-access-znr77\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043003 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll"] Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.043317 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043336 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.043361 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="extract-content" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043372 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="extract-content" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.043389 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="registry-server" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043402 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="registry-server" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.043417 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="extract-utilities" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043427 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="extract-utilities" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043586 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerName="registry-server" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.043609 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerName="console" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.044542 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.046621 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.056044 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll"] Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070164 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mg4h6_c5794771-53a6-46dd-bd10-b37f6534b87b/console/0.log" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070209 4837 generic.go:334] "Generic (PLEG): container finished" podID="c5794771-53a6-46dd-bd10-b37f6534b87b" containerID="1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc" exitCode=2 Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mg4h6" event={"ID":"c5794771-53a6-46dd-bd10-b37f6534b87b","Type":"ContainerDied","Data":"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc"} Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070287 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mg4h6" event={"ID":"c5794771-53a6-46dd-bd10-b37f6534b87b","Type":"ContainerDied","Data":"87750c41e5bf324c1aebdfaa153201e1b15e7a0f4b66af6f2922128d7396d607"} Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070304 4837 scope.go:117] "RemoveContainer" containerID="1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.070637 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mg4h6" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.073609 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzf77" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.073538 4837 generic.go:334] "Generic (PLEG): container finished" podID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" containerID="521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30" exitCode=0 Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.073643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerDied","Data":"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30"} Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.073713 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzf77" event={"ID":"a471a5ad-fe42-42a1-b2d6-3958fad91f5c","Type":"ContainerDied","Data":"d72aac1e62c282a4b356ce484bfcebc928992d5430312979b08b4a24bfb71327"} Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.092914 4837 scope.go:117] "RemoveContainer" containerID="1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.093509 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc\": container with ID starting with 1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc not found: ID does not exist" containerID="1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.093566 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc"} err="failed to get container status \"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc\": rpc error: code = NotFound desc = could not find container \"1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc\": container with ID starting with 1a2a3576e789673945a2899cb97614ce63e8e6f8cf2fc24c7f7bec70d95a41cc not found: ID does not exist" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.093597 4837 scope.go:117] "RemoveContainer" containerID="521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.100056 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.105828 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mg4h6"] Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.110255 4837 scope.go:117] "RemoveContainer" containerID="efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.121715 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.126035 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qzf77"] Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.133884 4837 scope.go:117] "RemoveContainer" containerID="69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.150379 4837 scope.go:117] "RemoveContainer" containerID="521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.150983 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30\": container with ID starting with 521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30 not found: ID does not exist" containerID="521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.151039 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30"} err="failed to get container status \"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30\": rpc error: code = NotFound desc = could not find container \"521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30\": container with ID starting with 521b78280c7b73d8e8947e6fbd21dea7ed3a8004659b4173d62f987529026d30 not found: ID does not exist" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.151072 4837 scope.go:117] "RemoveContainer" containerID="efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.151379 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e\": container with ID starting with efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e not found: ID does not exist" containerID="efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.151495 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e"} err="failed to get container status \"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e\": rpc error: code = NotFound desc = could not find container \"efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e\": container with ID starting with efeed4d48ac52c798f5fbb1e59b2edab7de07f3bc24897c3454248c7afd7001e not found: ID does not exist" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.151587 4837 scope.go:117] "RemoveContainer" containerID="69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921" Oct 01 07:19:50 crc kubenswrapper[4837]: E1001 07:19:50.151924 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921\": container with ID starting with 69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921 not found: ID does not exist" containerID="69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.152023 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921"} err="failed to get container status \"69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921\": rpc error: code = NotFound desc = could not find container \"69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921\": container with ID starting with 69c0e939f43779e04d17c40487727957d3e1fabfbd07f86a2e73a15ef8ca1921 not found: ID does not exist" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.221138 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.221419 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.221643 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkkf8\" (UniqueName: \"kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.322751 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkkf8\" (UniqueName: \"kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.322926 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.323005 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.324202 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.324909 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.359546 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkkf8\" (UniqueName: \"kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.370773 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:50 crc kubenswrapper[4837]: I1001 07:19:50.654074 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll"] Oct 01 07:19:51 crc kubenswrapper[4837]: I1001 07:19:51.088135 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerID="b1ede92fce966d89a53932e4b996c67cc70d0bdc42615853b43753228ffa3ff3" exitCode=0 Oct 01 07:19:51 crc kubenswrapper[4837]: I1001 07:19:51.088196 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" event={"ID":"c2c30e8f-8087-40d0-845a-4ea48e0547c6","Type":"ContainerDied","Data":"b1ede92fce966d89a53932e4b996c67cc70d0bdc42615853b43753228ffa3ff3"} Oct 01 07:19:51 crc kubenswrapper[4837]: I1001 07:19:51.088237 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" event={"ID":"c2c30e8f-8087-40d0-845a-4ea48e0547c6","Type":"ContainerStarted","Data":"c9e3354096f8908e6ef9cc9f9d01af700a8ea5f28eca3a8d84f404155fb0052a"} Oct 01 07:19:51 crc kubenswrapper[4837]: I1001 07:19:51.827030 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a471a5ad-fe42-42a1-b2d6-3958fad91f5c" path="/var/lib/kubelet/pods/a471a5ad-fe42-42a1-b2d6-3958fad91f5c/volumes" Oct 01 07:19:51 crc kubenswrapper[4837]: I1001 07:19:51.828427 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5794771-53a6-46dd-bd10-b37f6534b87b" path="/var/lib/kubelet/pods/c5794771-53a6-46dd-bd10-b37f6534b87b/volumes" Oct 01 07:19:53 crc kubenswrapper[4837]: I1001 07:19:53.108976 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerID="467f5c36581523dcad512963b05546cb1515c33f4cc2d93a9c6fddc3331a4b24" exitCode=0 Oct 01 07:19:53 crc kubenswrapper[4837]: I1001 07:19:53.109046 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" event={"ID":"c2c30e8f-8087-40d0-845a-4ea48e0547c6","Type":"ContainerDied","Data":"467f5c36581523dcad512963b05546cb1515c33f4cc2d93a9c6fddc3331a4b24"} Oct 01 07:19:54 crc kubenswrapper[4837]: I1001 07:19:54.120787 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerID="d9ad2d7a844fbd16b30420b6e222843c310d48259e7f3bd37113d3ffd154b1c1" exitCode=0 Oct 01 07:19:54 crc kubenswrapper[4837]: I1001 07:19:54.120866 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" event={"ID":"c2c30e8f-8087-40d0-845a-4ea48e0547c6","Type":"ContainerDied","Data":"d9ad2d7a844fbd16b30420b6e222843c310d48259e7f3bd37113d3ffd154b1c1"} Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.490238 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.637795 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util\") pod \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.637993 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle\") pod \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.638067 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkkf8\" (UniqueName: \"kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8\") pod \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\" (UID: \"c2c30e8f-8087-40d0-845a-4ea48e0547c6\") " Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.641456 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle" (OuterVolumeSpecName: "bundle") pod "c2c30e8f-8087-40d0-845a-4ea48e0547c6" (UID: "c2c30e8f-8087-40d0-845a-4ea48e0547c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.649849 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8" (OuterVolumeSpecName: "kube-api-access-jkkf8") pod "c2c30e8f-8087-40d0-845a-4ea48e0547c6" (UID: "c2c30e8f-8087-40d0-845a-4ea48e0547c6"). InnerVolumeSpecName "kube-api-access-jkkf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.661990 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util" (OuterVolumeSpecName: "util") pod "c2c30e8f-8087-40d0-845a-4ea48e0547c6" (UID: "c2c30e8f-8087-40d0-845a-4ea48e0547c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.739873 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkkf8\" (UniqueName: \"kubernetes.io/projected/c2c30e8f-8087-40d0-845a-4ea48e0547c6-kube-api-access-jkkf8\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.739922 4837 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-util\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:55 crc kubenswrapper[4837]: I1001 07:19:55.739942 4837 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c2c30e8f-8087-40d0-845a-4ea48e0547c6-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:19:56 crc kubenswrapper[4837]: I1001 07:19:56.143320 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" event={"ID":"c2c30e8f-8087-40d0-845a-4ea48e0547c6","Type":"ContainerDied","Data":"c9e3354096f8908e6ef9cc9f9d01af700a8ea5f28eca3a8d84f404155fb0052a"} Oct 01 07:19:56 crc kubenswrapper[4837]: I1001 07:19:56.143382 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9e3354096f8908e6ef9cc9f9d01af700a8ea5f28eca3a8d84f404155fb0052a" Oct 01 07:19:56 crc kubenswrapper[4837]: I1001 07:19:56.143416 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.892434 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv"] Oct 01 07:20:05 crc kubenswrapper[4837]: E1001 07:20:05.893160 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="util" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.893174 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="util" Oct 01 07:20:05 crc kubenswrapper[4837]: E1001 07:20:05.893185 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="extract" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.893194 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="extract" Oct 01 07:20:05 crc kubenswrapper[4837]: E1001 07:20:05.893211 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="pull" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.893220 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="pull" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.893349 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c30e8f-8087-40d0-845a-4ea48e0547c6" containerName="extract" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.893780 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.895639 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.895864 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.896260 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-wnsk7" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.896408 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.896640 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.907978 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv"] Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.986137 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-apiservice-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.986244 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-webhook-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:05 crc kubenswrapper[4837]: I1001 07:20:05.986313 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpcqc\" (UniqueName: \"kubernetes.io/projected/b06a7555-980f-454d-8459-7385bcdd4471-kube-api-access-gpcqc\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.087340 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-apiservice-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.087866 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-webhook-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.087935 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpcqc\" (UniqueName: \"kubernetes.io/projected/b06a7555-980f-454d-8459-7385bcdd4471-kube-api-access-gpcqc\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.094476 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-webhook-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.094941 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b06a7555-980f-454d-8459-7385bcdd4471-apiservice-cert\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.106323 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpcqc\" (UniqueName: \"kubernetes.io/projected/b06a7555-980f-454d-8459-7385bcdd4471-kube-api-access-gpcqc\") pod \"metallb-operator-controller-manager-75b5d968bd-ggvqv\" (UID: \"b06a7555-980f-454d-8459-7385bcdd4471\") " pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.128892 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-876c594b8-snjgm"] Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.129491 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.131475 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.131739 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.131918 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-t5wt5" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.143880 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-876c594b8-snjgm"] Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.189247 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-apiservice-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.189341 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhst\" (UniqueName: \"kubernetes.io/projected/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-kube-api-access-fwhst\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.189369 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-webhook-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.207801 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.289920 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-apiservice-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.289976 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhst\" (UniqueName: \"kubernetes.io/projected/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-kube-api-access-fwhst\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.290003 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-webhook-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.301295 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-webhook-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.310933 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-apiservice-cert\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.314408 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhst\" (UniqueName: \"kubernetes.io/projected/a2da66dd-d078-41cb-ae82-0aa30cdc04cd-kube-api-access-fwhst\") pod \"metallb-operator-webhook-server-876c594b8-snjgm\" (UID: \"a2da66dd-d078-41cb-ae82-0aa30cdc04cd\") " pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.443777 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.673517 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv"] Oct 01 07:20:06 crc kubenswrapper[4837]: W1001 07:20:06.678196 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb06a7555_980f_454d_8459_7385bcdd4471.slice/crio-326df9b57d3c2f96f8ebfa829449e1d0f09ea26d5f538d9617be85c371753b62 WatchSource:0}: Error finding container 326df9b57d3c2f96f8ebfa829449e1d0f09ea26d5f538d9617be85c371753b62: Status 404 returned error can't find the container with id 326df9b57d3c2f96f8ebfa829449e1d0f09ea26d5f538d9617be85c371753b62 Oct 01 07:20:06 crc kubenswrapper[4837]: I1001 07:20:06.761433 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-876c594b8-snjgm"] Oct 01 07:20:06 crc kubenswrapper[4837]: W1001 07:20:06.768884 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2da66dd_d078_41cb_ae82_0aa30cdc04cd.slice/crio-29ddf08c26804e9799bbe81f6a2bf23fb3885e966ce83eea61827e694b0ddae6 WatchSource:0}: Error finding container 29ddf08c26804e9799bbe81f6a2bf23fb3885e966ce83eea61827e694b0ddae6: Status 404 returned error can't find the container with id 29ddf08c26804e9799bbe81f6a2bf23fb3885e966ce83eea61827e694b0ddae6 Oct 01 07:20:07 crc kubenswrapper[4837]: I1001 07:20:07.213326 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" event={"ID":"a2da66dd-d078-41cb-ae82-0aa30cdc04cd","Type":"ContainerStarted","Data":"29ddf08c26804e9799bbe81f6a2bf23fb3885e966ce83eea61827e694b0ddae6"} Oct 01 07:20:07 crc kubenswrapper[4837]: I1001 07:20:07.214937 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" event={"ID":"b06a7555-980f-454d-8459-7385bcdd4471","Type":"ContainerStarted","Data":"326df9b57d3c2f96f8ebfa829449e1d0f09ea26d5f538d9617be85c371753b62"} Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.249466 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" event={"ID":"b06a7555-980f-454d-8459-7385bcdd4471","Type":"ContainerStarted","Data":"f3c66c8024393c3e3cc820f586b4d05fd42d52a01a582947dd4aca907cc39dcd"} Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.250096 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.251992 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" event={"ID":"a2da66dd-d078-41cb-ae82-0aa30cdc04cd","Type":"ContainerStarted","Data":"9bf3eef2467aa1d4d41391fe6783cce4cd97a015092c7cd655cce0129f14b936"} Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.252151 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.280680 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" podStartSLOduration=2.082587195 podStartE2EDuration="7.280663896s" podCreationTimestamp="2025-10-01 07:20:05 +0000 UTC" firstStartedPulling="2025-10-01 07:20:06.680019967 +0000 UTC m=+863.521627422" lastFinishedPulling="2025-10-01 07:20:11.878096678 +0000 UTC m=+868.719704123" observedRunningTime="2025-10-01 07:20:12.277965371 +0000 UTC m=+869.119572826" watchObservedRunningTime="2025-10-01 07:20:12.280663896 +0000 UTC m=+869.122271351" Oct 01 07:20:12 crc kubenswrapper[4837]: I1001 07:20:12.309163 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" podStartSLOduration=1.177446673 podStartE2EDuration="6.309145267s" podCreationTimestamp="2025-10-01 07:20:06 +0000 UTC" firstStartedPulling="2025-10-01 07:20:06.771394911 +0000 UTC m=+863.613002356" lastFinishedPulling="2025-10-01 07:20:11.903093495 +0000 UTC m=+868.744700950" observedRunningTime="2025-10-01 07:20:12.306003641 +0000 UTC m=+869.147611106" watchObservedRunningTime="2025-10-01 07:20:12.309145267 +0000 UTC m=+869.150752722" Oct 01 07:20:23 crc kubenswrapper[4837]: I1001 07:20:23.079317 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:20:23 crc kubenswrapper[4837]: I1001 07:20:23.079841 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:20:26 crc kubenswrapper[4837]: I1001 07:20:26.453131 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-876c594b8-snjgm" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.212180 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-75b5d968bd-ggvqv" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.971603 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr"] Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.972477 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.977934 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-26vm7" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.978020 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.981228 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-qr2xb"] Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.986263 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr"] Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.986412 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.990339 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 01 07:20:46 crc kubenswrapper[4837]: I1001 07:20:46.990605 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.058152 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-c4jff"] Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.059139 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.060463 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.062223 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.062373 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-c4hwb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.062493 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.072541 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-hdqz9"] Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.073665 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.077013 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-hdqz9"] Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.078572 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104211 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78113597-c462-4e42-8439-6d8b74465dbf-cert\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104261 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-conf\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104300 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104313 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-startup\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104331 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-reloader\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104352 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-sockets\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104379 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics-certs\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104399 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdncs\" (UniqueName: \"kubernetes.io/projected/78113597-c462-4e42-8439-6d8b74465dbf-kube-api-access-fdncs\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.104426 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2wd6\" (UniqueName: \"kubernetes.io/projected/25ceb3dd-39ef-4039-b923-da36fb01ff05-kube-api-access-h2wd6\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205449 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78113597-c462-4e42-8439-6d8b74465dbf-cert\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205493 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205510 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-conf\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205538 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-cert\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205568 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-startup\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205583 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metallb-excludel2\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205618 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75vdd\" (UniqueName: \"kubernetes.io/projected/116f7e17-f8af-4518-af7d-f2ad1ab06acf-kube-api-access-75vdd\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205638 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-reloader\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205658 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-sockets\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205703 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics-certs\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205720 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205753 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-metrics-certs\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205770 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdncs\" (UniqueName: \"kubernetes.io/projected/78113597-c462-4e42-8439-6d8b74465dbf-kube-api-access-fdncs\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205789 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mftwt\" (UniqueName: \"kubernetes.io/projected/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-kube-api-access-mftwt\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.205815 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2wd6\" (UniqueName: \"kubernetes.io/projected/25ceb3dd-39ef-4039-b923-da36fb01ff05-kube-api-access-h2wd6\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.206394 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-conf\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.206732 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-sockets\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.206882 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-reloader\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.207147 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.208403 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/25ceb3dd-39ef-4039-b923-da36fb01ff05-frr-startup\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.215399 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25ceb3dd-39ef-4039-b923-da36fb01ff05-metrics-certs\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.215665 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/78113597-c462-4e42-8439-6d8b74465dbf-cert\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.222070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2wd6\" (UniqueName: \"kubernetes.io/projected/25ceb3dd-39ef-4039-b923-da36fb01ff05-kube-api-access-h2wd6\") pod \"frr-k8s-qr2xb\" (UID: \"25ceb3dd-39ef-4039-b923-da36fb01ff05\") " pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.228435 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdncs\" (UniqueName: \"kubernetes.io/projected/78113597-c462-4e42-8439-6d8b74465dbf-kube-api-access-fdncs\") pod \"frr-k8s-webhook-server-5478bdb765-s9dpr\" (UID: \"78113597-c462-4e42-8439-6d8b74465dbf\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.287954 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.303055 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.306806 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.306884 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-cert\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.306944 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metallb-excludel2\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.306980 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75vdd\" (UniqueName: \"kubernetes.io/projected/116f7e17-f8af-4518-af7d-f2ad1ab06acf-kube-api-access-75vdd\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.307067 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.307106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-metrics-certs\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.307140 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mftwt\" (UniqueName: \"kubernetes.io/projected/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-kube-api-access-mftwt\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.307661 4837 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.307763 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs podName:96e8ea69-4a48-4923-94c7-cd3fb01b6f69 nodeName:}" failed. No retries permitted until 2025-10-01 07:20:47.8077396 +0000 UTC m=+904.649347095 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs") pod "speaker-c4jff" (UID: "96e8ea69-4a48-4923-94c7-cd3fb01b6f69") : secret "speaker-certs-secret" not found Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.309337 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metallb-excludel2\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.310494 4837 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.310563 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist podName:96e8ea69-4a48-4923-94c7-cd3fb01b6f69 nodeName:}" failed. No retries permitted until 2025-10-01 07:20:47.810545497 +0000 UTC m=+904.652152992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist") pod "speaker-c4jff" (UID: "96e8ea69-4a48-4923-94c7-cd3fb01b6f69") : secret "metallb-memberlist" not found Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.312546 4837 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.313581 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-metrics-certs\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.327411 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/116f7e17-f8af-4518-af7d-f2ad1ab06acf-cert\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.327660 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mftwt\" (UniqueName: \"kubernetes.io/projected/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-kube-api-access-mftwt\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.349386 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75vdd\" (UniqueName: \"kubernetes.io/projected/116f7e17-f8af-4518-af7d-f2ad1ab06acf-kube-api-access-75vdd\") pod \"controller-5d688f5ffc-hdqz9\" (UID: \"116f7e17-f8af-4518-af7d-f2ad1ab06acf\") " pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.417076 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.480924 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"2cacbaf90a2ab671749fa0e0eebf70bc43b2588ac06f1d054b6903d953a97f41"} Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.616542 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-hdqz9"] Oct 01 07:20:47 crc kubenswrapper[4837]: W1001 07:20:47.620897 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod116f7e17_f8af_4518_af7d_f2ad1ab06acf.slice/crio-ef1eae4dab8f157a12ec97af4b90d570a4fd0736cbe3fce27f61e578d65501bc WatchSource:0}: Error finding container ef1eae4dab8f157a12ec97af4b90d570a4fd0736cbe3fce27f61e578d65501bc: Status 404 returned error can't find the container with id ef1eae4dab8f157a12ec97af4b90d570a4fd0736cbe3fce27f61e578d65501bc Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.788536 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr"] Oct 01 07:20:47 crc kubenswrapper[4837]: W1001 07:20:47.795246 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78113597_c462_4e42_8439_6d8b74465dbf.slice/crio-550271704d743e3c09e0fdfdd91330c218a486d9d209438182cb8d9321e20d59 WatchSource:0}: Error finding container 550271704d743e3c09e0fdfdd91330c218a486d9d209438182cb8d9321e20d59: Status 404 returned error can't find the container with id 550271704d743e3c09e0fdfdd91330c218a486d9d209438182cb8d9321e20d59 Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.815529 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.815811 4837 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 01 07:20:47 crc kubenswrapper[4837]: E1001 07:20:47.815881 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist podName:96e8ea69-4a48-4923-94c7-cd3fb01b6f69 nodeName:}" failed. No retries permitted until 2025-10-01 07:20:48.815863971 +0000 UTC m=+905.657471436 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist") pod "speaker-c4jff" (UID: "96e8ea69-4a48-4923-94c7-cd3fb01b6f69") : secret "metallb-memberlist" not found Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.816276 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:47 crc kubenswrapper[4837]: I1001 07:20:47.830379 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-metrics-certs\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.490917 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-hdqz9" event={"ID":"116f7e17-f8af-4518-af7d-f2ad1ab06acf","Type":"ContainerStarted","Data":"1953b86b74c44c705a3f9f5f642d2f68e0e2aaa27a1480f2dbb894d58f72d45b"} Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.491452 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.491483 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-hdqz9" event={"ID":"116f7e17-f8af-4518-af7d-f2ad1ab06acf","Type":"ContainerStarted","Data":"d36e4a62417bc15c7e23421c7187fc6c89684255abaef68cd4943b78a5e69c20"} Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.491510 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-hdqz9" event={"ID":"116f7e17-f8af-4518-af7d-f2ad1ab06acf","Type":"ContainerStarted","Data":"ef1eae4dab8f157a12ec97af4b90d570a4fd0736cbe3fce27f61e578d65501bc"} Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.492577 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" event={"ID":"78113597-c462-4e42-8439-6d8b74465dbf","Type":"ContainerStarted","Data":"550271704d743e3c09e0fdfdd91330c218a486d9d209438182cb8d9321e20d59"} Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.527642 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-hdqz9" podStartSLOduration=1.5276172350000001 podStartE2EDuration="1.527617235s" podCreationTimestamp="2025-10-01 07:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:20:48.519359102 +0000 UTC m=+905.360966597" watchObservedRunningTime="2025-10-01 07:20:48.527617235 +0000 UTC m=+905.369224720" Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.831455 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.837821 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/96e8ea69-4a48-4923-94c7-cd3fb01b6f69-memberlist\") pod \"speaker-c4jff\" (UID: \"96e8ea69-4a48-4923-94c7-cd3fb01b6f69\") " pod="metallb-system/speaker-c4jff" Oct 01 07:20:48 crc kubenswrapper[4837]: I1001 07:20:48.870861 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-c4jff" Oct 01 07:20:48 crc kubenswrapper[4837]: W1001 07:20:48.909796 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96e8ea69_4a48_4923_94c7_cd3fb01b6f69.slice/crio-608eb5aaeaf407365fa409943d220e365a1a6d5440f8b62bfdd4e84bf45d2a99 WatchSource:0}: Error finding container 608eb5aaeaf407365fa409943d220e365a1a6d5440f8b62bfdd4e84bf45d2a99: Status 404 returned error can't find the container with id 608eb5aaeaf407365fa409943d220e365a1a6d5440f8b62bfdd4e84bf45d2a99 Oct 01 07:20:49 crc kubenswrapper[4837]: I1001 07:20:49.521306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c4jff" event={"ID":"96e8ea69-4a48-4923-94c7-cd3fb01b6f69","Type":"ContainerStarted","Data":"c039cecdde7bca47248b8ec10d79cf11dc7da7f135ac62d027fd4a35b162e007"} Oct 01 07:20:49 crc kubenswrapper[4837]: I1001 07:20:49.521352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c4jff" event={"ID":"96e8ea69-4a48-4923-94c7-cd3fb01b6f69","Type":"ContainerStarted","Data":"608eb5aaeaf407365fa409943d220e365a1a6d5440f8b62bfdd4e84bf45d2a99"} Oct 01 07:20:50 crc kubenswrapper[4837]: I1001 07:20:50.529918 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c4jff" event={"ID":"96e8ea69-4a48-4923-94c7-cd3fb01b6f69","Type":"ContainerStarted","Data":"2e40f046a27486c3891b83b5f58016191ebe439ef40b0445dcd96ebec4e6ffba"} Oct 01 07:20:50 crc kubenswrapper[4837]: I1001 07:20:50.530457 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-c4jff" Oct 01 07:20:50 crc kubenswrapper[4837]: I1001 07:20:50.550143 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-c4jff" podStartSLOduration=3.550117422 podStartE2EDuration="3.550117422s" podCreationTimestamp="2025-10-01 07:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:20:50.546495182 +0000 UTC m=+907.388102637" watchObservedRunningTime="2025-10-01 07:20:50.550117422 +0000 UTC m=+907.391724877" Oct 01 07:20:53 crc kubenswrapper[4837]: I1001 07:20:53.079314 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:20:53 crc kubenswrapper[4837]: I1001 07:20:53.079423 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:20:55 crc kubenswrapper[4837]: I1001 07:20:55.569802 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" event={"ID":"78113597-c462-4e42-8439-6d8b74465dbf","Type":"ContainerStarted","Data":"180c64b7eea4fcd8035d395a7f0d8f76db3b26bf0e4c28d8a208758d0b400478"} Oct 01 07:20:55 crc kubenswrapper[4837]: I1001 07:20:55.570825 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:20:55 crc kubenswrapper[4837]: I1001 07:20:55.572214 4837 generic.go:334] "Generic (PLEG): container finished" podID="25ceb3dd-39ef-4039-b923-da36fb01ff05" containerID="4f82623a3f65a589a1439f74e1c3a7fa3038bc8aa7977ff997e109f8a1e21ece" exitCode=0 Oct 01 07:20:55 crc kubenswrapper[4837]: I1001 07:20:55.572273 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerDied","Data":"4f82623a3f65a589a1439f74e1c3a7fa3038bc8aa7977ff997e109f8a1e21ece"} Oct 01 07:20:55 crc kubenswrapper[4837]: I1001 07:20:55.595662 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" podStartSLOduration=2.369098467 podStartE2EDuration="9.595631994s" podCreationTimestamp="2025-10-01 07:20:46 +0000 UTC" firstStartedPulling="2025-10-01 07:20:47.798086776 +0000 UTC m=+904.639694251" lastFinishedPulling="2025-10-01 07:20:55.024620313 +0000 UTC m=+911.866227778" observedRunningTime="2025-10-01 07:20:55.593590454 +0000 UTC m=+912.435197959" watchObservedRunningTime="2025-10-01 07:20:55.595631994 +0000 UTC m=+912.437239479" Oct 01 07:20:56 crc kubenswrapper[4837]: I1001 07:20:56.581955 4837 generic.go:334] "Generic (PLEG): container finished" podID="25ceb3dd-39ef-4039-b923-da36fb01ff05" containerID="d4922dc833ba7ed647c69425a91e76f368f9ece2b3a4def1c99a0e13cd7fecff" exitCode=0 Oct 01 07:20:56 crc kubenswrapper[4837]: I1001 07:20:56.582032 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerDied","Data":"d4922dc833ba7ed647c69425a91e76f368f9ece2b3a4def1c99a0e13cd7fecff"} Oct 01 07:20:57 crc kubenswrapper[4837]: I1001 07:20:57.426154 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-hdqz9" Oct 01 07:20:57 crc kubenswrapper[4837]: I1001 07:20:57.592153 4837 generic.go:334] "Generic (PLEG): container finished" podID="25ceb3dd-39ef-4039-b923-da36fb01ff05" containerID="b6523a887d95fb150764b9c901a59f0da4cb54c79ccc38d434577f2d0a13b869" exitCode=0 Oct 01 07:20:57 crc kubenswrapper[4837]: I1001 07:20:57.592218 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerDied","Data":"b6523a887d95fb150764b9c901a59f0da4cb54c79ccc38d434577f2d0a13b869"} Oct 01 07:20:58 crc kubenswrapper[4837]: I1001 07:20:58.612483 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"b2e0a1737e9e0321995f64ac8c8ce1fea159fa8bb6f41ffab62c321481b08933"} Oct 01 07:20:58 crc kubenswrapper[4837]: I1001 07:20:58.612813 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"d924da8b03ebc03df3f8da7135e1ea92c331fef65f99bcb544c8dfeca075ac2d"} Oct 01 07:20:58 crc kubenswrapper[4837]: I1001 07:20:58.612825 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"0a067582f790113223f33aaff692ac019ebeae4b8fe8c9da33fea597cfb133cd"} Oct 01 07:20:58 crc kubenswrapper[4837]: I1001 07:20:58.612835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"48e254f9e1e891e111fe91be00e113148fda8eb85799833a51a3d3f4b6b49d47"} Oct 01 07:20:58 crc kubenswrapper[4837]: I1001 07:20:58.612843 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"d1d0c37afb1bac0beb1fbb46cff50e588b48bee67f182a728561eb23558d6398"} Oct 01 07:20:59 crc kubenswrapper[4837]: I1001 07:20:59.627925 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qr2xb" event={"ID":"25ceb3dd-39ef-4039-b923-da36fb01ff05","Type":"ContainerStarted","Data":"b5e120d527330e13f5116d45c72099e043f7f04bdaf15cff89efe2251dbe40b7"} Oct 01 07:20:59 crc kubenswrapper[4837]: I1001 07:20:59.628953 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:20:59 crc kubenswrapper[4837]: I1001 07:20:59.673238 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-qr2xb" podStartSLOduration=6.161722587 podStartE2EDuration="13.673215492s" podCreationTimestamp="2025-10-01 07:20:46 +0000 UTC" firstStartedPulling="2025-10-01 07:20:47.464042765 +0000 UTC m=+904.305650220" lastFinishedPulling="2025-10-01 07:20:54.97553566 +0000 UTC m=+911.817143125" observedRunningTime="2025-10-01 07:20:59.670817463 +0000 UTC m=+916.512424978" watchObservedRunningTime="2025-10-01 07:20:59.673215492 +0000 UTC m=+916.514822987" Oct 01 07:21:02 crc kubenswrapper[4837]: I1001 07:21:02.304444 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:21:02 crc kubenswrapper[4837]: I1001 07:21:02.382112 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:21:07 crc kubenswrapper[4837]: I1001 07:21:07.295426 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-s9dpr" Oct 01 07:21:07 crc kubenswrapper[4837]: I1001 07:21:07.309298 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-qr2xb" Oct 01 07:21:08 crc kubenswrapper[4837]: I1001 07:21:08.877353 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-c4jff" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.399308 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4"] Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.401609 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.405872 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.415406 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4"] Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.471380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.471466 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56gkk\" (UniqueName: \"kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.471554 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.573318 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.573406 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.573432 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56gkk\" (UniqueName: \"kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.574229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.574268 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.601411 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56gkk\" (UniqueName: \"kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:10 crc kubenswrapper[4837]: I1001 07:21:10.731763 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:11 crc kubenswrapper[4837]: I1001 07:21:11.253545 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4"] Oct 01 07:21:11 crc kubenswrapper[4837]: W1001 07:21:11.259721 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8c13b9c_27b5_44bb_b820_f1e4dca3e302.slice/crio-4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59 WatchSource:0}: Error finding container 4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59: Status 404 returned error can't find the container with id 4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59 Oct 01 07:21:11 crc kubenswrapper[4837]: I1001 07:21:11.729820 4837 generic.go:334] "Generic (PLEG): container finished" podID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerID="40826f0aebc0b3c822beaeaddf84f0a24bcff0cafb9df077304c18580b364a08" exitCode=0 Oct 01 07:21:11 crc kubenswrapper[4837]: I1001 07:21:11.730325 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" event={"ID":"f8c13b9c-27b5-44bb-b820-f1e4dca3e302","Type":"ContainerDied","Data":"40826f0aebc0b3c822beaeaddf84f0a24bcff0cafb9df077304c18580b364a08"} Oct 01 07:21:11 crc kubenswrapper[4837]: I1001 07:21:11.730375 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" event={"ID":"f8c13b9c-27b5-44bb-b820-f1e4dca3e302","Type":"ContainerStarted","Data":"4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59"} Oct 01 07:21:15 crc kubenswrapper[4837]: I1001 07:21:15.760751 4837 generic.go:334] "Generic (PLEG): container finished" podID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerID="712f7433a7ff662b358672115beceeb1740eeb036fb27a385de78fb2337062d8" exitCode=0 Oct 01 07:21:15 crc kubenswrapper[4837]: I1001 07:21:15.760830 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" event={"ID":"f8c13b9c-27b5-44bb-b820-f1e4dca3e302","Type":"ContainerDied","Data":"712f7433a7ff662b358672115beceeb1740eeb036fb27a385de78fb2337062d8"} Oct 01 07:21:16 crc kubenswrapper[4837]: I1001 07:21:16.774181 4837 generic.go:334] "Generic (PLEG): container finished" podID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerID="9f7390665ad11ccf9c59cccc8b6a8b9cec05e4c322f69089b0d484849c1e606f" exitCode=0 Oct 01 07:21:16 crc kubenswrapper[4837]: I1001 07:21:16.774277 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" event={"ID":"f8c13b9c-27b5-44bb-b820-f1e4dca3e302","Type":"ContainerDied","Data":"9f7390665ad11ccf9c59cccc8b6a8b9cec05e4c322f69089b0d484849c1e606f"} Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.192563 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.280040 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle\") pod \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.280115 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util\") pod \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.280179 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56gkk\" (UniqueName: \"kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk\") pod \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\" (UID: \"f8c13b9c-27b5-44bb-b820-f1e4dca3e302\") " Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.281761 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle" (OuterVolumeSpecName: "bundle") pod "f8c13b9c-27b5-44bb-b820-f1e4dca3e302" (UID: "f8c13b9c-27b5-44bb-b820-f1e4dca3e302"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.291898 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk" (OuterVolumeSpecName: "kube-api-access-56gkk") pod "f8c13b9c-27b5-44bb-b820-f1e4dca3e302" (UID: "f8c13b9c-27b5-44bb-b820-f1e4dca3e302"). InnerVolumeSpecName "kube-api-access-56gkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.301658 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util" (OuterVolumeSpecName: "util") pod "f8c13b9c-27b5-44bb-b820-f1e4dca3e302" (UID: "f8c13b9c-27b5-44bb-b820-f1e4dca3e302"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.381518 4837 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.381570 4837 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-util\") on node \"crc\" DevicePath \"\"" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.381590 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56gkk\" (UniqueName: \"kubernetes.io/projected/f8c13b9c-27b5-44bb-b820-f1e4dca3e302-kube-api-access-56gkk\") on node \"crc\" DevicePath \"\"" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.795821 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" event={"ID":"f8c13b9c-27b5-44bb-b820-f1e4dca3e302","Type":"ContainerDied","Data":"4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59"} Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.795878 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4133f80e1e94b1306d83b8fd39063007ee0898b8782f765d23b5e82b4b5c8c59" Oct 01 07:21:18 crc kubenswrapper[4837]: I1001 07:21:18.795919 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.079252 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.079791 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.079836 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.080361 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.080413 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb" gracePeriod=600 Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.262662 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw"] Oct 01 07:21:23 crc kubenswrapper[4837]: E1001 07:21:23.262880 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="extract" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.262891 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="extract" Oct 01 07:21:23 crc kubenswrapper[4837]: E1001 07:21:23.262905 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="pull" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.262912 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="pull" Oct 01 07:21:23 crc kubenswrapper[4837]: E1001 07:21:23.262934 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="util" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.262940 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="util" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.263033 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c13b9c-27b5-44bb-b820-f1e4dca3e302" containerName="extract" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.263382 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.266661 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.266725 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.266919 4837 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-9gpf6" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.292890 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw"] Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.343941 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5rkl\" (UniqueName: \"kubernetes.io/projected/61439274-057f-4808-a120-72dfe0f354fe-kube-api-access-t5rkl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-l2zxw\" (UID: \"61439274-057f-4808-a120-72dfe0f354fe\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.445463 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5rkl\" (UniqueName: \"kubernetes.io/projected/61439274-057f-4808-a120-72dfe0f354fe-kube-api-access-t5rkl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-l2zxw\" (UID: \"61439274-057f-4808-a120-72dfe0f354fe\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.468969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5rkl\" (UniqueName: \"kubernetes.io/projected/61439274-057f-4808-a120-72dfe0f354fe-kube-api-access-t5rkl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-l2zxw\" (UID: \"61439274-057f-4808-a120-72dfe0f354fe\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.577284 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.836321 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb" exitCode=0 Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.836599 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb"} Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.836624 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86"} Oct 01 07:21:23 crc kubenswrapper[4837]: I1001 07:21:23.836815 4837 scope.go:117] "RemoveContainer" containerID="875a344eadd19f387f54a3997c44687922a9d3b1f54aaaec33d3f870e219fa05" Oct 01 07:21:24 crc kubenswrapper[4837]: I1001 07:21:24.052976 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw"] Oct 01 07:21:24 crc kubenswrapper[4837]: W1001 07:21:24.060007 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61439274_057f_4808_a120_72dfe0f354fe.slice/crio-afa4d62809ed1fe92b4a459be97a0ae9e6bbee3bf2d36ad30cb93033b17bb1cd WatchSource:0}: Error finding container afa4d62809ed1fe92b4a459be97a0ae9e6bbee3bf2d36ad30cb93033b17bb1cd: Status 404 returned error can't find the container with id afa4d62809ed1fe92b4a459be97a0ae9e6bbee3bf2d36ad30cb93033b17bb1cd Oct 01 07:21:24 crc kubenswrapper[4837]: I1001 07:21:24.846676 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" event={"ID":"61439274-057f-4808-a120-72dfe0f354fe","Type":"ContainerStarted","Data":"afa4d62809ed1fe92b4a459be97a0ae9e6bbee3bf2d36ad30cb93033b17bb1cd"} Oct 01 07:21:31 crc kubenswrapper[4837]: I1001 07:21:31.898037 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" event={"ID":"61439274-057f-4808-a120-72dfe0f354fe","Type":"ContainerStarted","Data":"7ee42cb249debe8d3ec43d1cc8ee994d68fbdd8c16b3653d9ad6fbfeb455548c"} Oct 01 07:21:31 crc kubenswrapper[4837]: I1001 07:21:31.919397 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-l2zxw" podStartSLOduration=2.137979168 podStartE2EDuration="8.919376288s" podCreationTimestamp="2025-10-01 07:21:23 +0000 UTC" firstStartedPulling="2025-10-01 07:21:24.062226382 +0000 UTC m=+940.903833837" lastFinishedPulling="2025-10-01 07:21:30.843623502 +0000 UTC m=+947.685230957" observedRunningTime="2025-10-01 07:21:31.916528777 +0000 UTC m=+948.758136262" watchObservedRunningTime="2025-10-01 07:21:31.919376288 +0000 UTC m=+948.760983773" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.772268 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-ktdtr"] Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.773399 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.775352 4837 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-9862h" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.775550 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.775967 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.812835 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-ktdtr"] Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.909246 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkg4h\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-kube-api-access-rkg4h\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:34 crc kubenswrapper[4837]: I1001 07:21:34.909420 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-bound-sa-token\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.010307 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-bound-sa-token\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.010387 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkg4h\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-kube-api-access-rkg4h\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.036864 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkg4h\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-kube-api-access-rkg4h\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.037733 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b345b2c-2564-4049-94b8-0ddc18b8bd74-bound-sa-token\") pod \"cert-manager-webhook-d969966f-ktdtr\" (UID: \"8b345b2c-2564-4049-94b8-0ddc18b8bd74\") " pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.089493 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.353240 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-ktdtr"] Oct 01 07:21:35 crc kubenswrapper[4837]: I1001 07:21:35.925183 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" event={"ID":"8b345b2c-2564-4049-94b8-0ddc18b8bd74","Type":"ContainerStarted","Data":"aaf2b7a76d6b7f9018f913778c14e1457a36aeb5175a004f0bf11b4fa8a9d4cc"} Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.001042 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb"] Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.001933 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.007815 4837 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b9n4z" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.011440 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb"] Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.047285 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxcjg\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-kube-api-access-wxcjg\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.047329 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.150842 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxcjg\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-kube-api-access-wxcjg\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.150887 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.175661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.178850 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxcjg\" (UniqueName: \"kubernetes.io/projected/262862ea-9ee9-4013-85ca-f8d8a196cd00-kube-api-access-wxcjg\") pod \"cert-manager-cainjector-7d9f95dbf-xxpxb\" (UID: \"262862ea-9ee9-4013-85ca-f8d8a196cd00\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.318163 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.745639 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb"] Oct 01 07:21:37 crc kubenswrapper[4837]: W1001 07:21:37.751404 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod262862ea_9ee9_4013_85ca_f8d8a196cd00.slice/crio-ba9421b776d1a010682c6b0c30cc237101e33fe407a42869fa78c22feae881f8 WatchSource:0}: Error finding container ba9421b776d1a010682c6b0c30cc237101e33fe407a42869fa78c22feae881f8: Status 404 returned error can't find the container with id ba9421b776d1a010682c6b0c30cc237101e33fe407a42869fa78c22feae881f8 Oct 01 07:21:37 crc kubenswrapper[4837]: I1001 07:21:37.934933 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" event={"ID":"262862ea-9ee9-4013-85ca-f8d8a196cd00","Type":"ContainerStarted","Data":"ba9421b776d1a010682c6b0c30cc237101e33fe407a42869fa78c22feae881f8"} Oct 01 07:21:39 crc kubenswrapper[4837]: I1001 07:21:39.951535 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" event={"ID":"262862ea-9ee9-4013-85ca-f8d8a196cd00","Type":"ContainerStarted","Data":"1e41c5d87995d9cf12551d3fd80b1c245fe5aa77ec53c97a1e7ba3cd7cfb1d7d"} Oct 01 07:21:39 crc kubenswrapper[4837]: I1001 07:21:39.953437 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" event={"ID":"8b345b2c-2564-4049-94b8-0ddc18b8bd74","Type":"ContainerStarted","Data":"32c1936c108c52840a48f80e13a69a13f6f668dcbb994e5ea8f1df4342b918cf"} Oct 01 07:21:39 crc kubenswrapper[4837]: I1001 07:21:39.954123 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:40 crc kubenswrapper[4837]: I1001 07:21:40.006667 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" podStartSLOduration=1.676152313 podStartE2EDuration="6.006651008s" podCreationTimestamp="2025-10-01 07:21:34 +0000 UTC" firstStartedPulling="2025-10-01 07:21:35.362744364 +0000 UTC m=+952.204351819" lastFinishedPulling="2025-10-01 07:21:39.693243059 +0000 UTC m=+956.534850514" observedRunningTime="2025-10-01 07:21:40.004528996 +0000 UTC m=+956.846136471" watchObservedRunningTime="2025-10-01 07:21:40.006651008 +0000 UTC m=+956.848258483" Oct 01 07:21:40 crc kubenswrapper[4837]: I1001 07:21:40.008558 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-xxpxb" podStartSLOduration=2.0682305579999998 podStartE2EDuration="4.008548155s" podCreationTimestamp="2025-10-01 07:21:36 +0000 UTC" firstStartedPulling="2025-10-01 07:21:37.753344352 +0000 UTC m=+954.594951807" lastFinishedPulling="2025-10-01 07:21:39.693661949 +0000 UTC m=+956.535269404" observedRunningTime="2025-10-01 07:21:39.975099979 +0000 UTC m=+956.816707474" watchObservedRunningTime="2025-10-01 07:21:40.008548155 +0000 UTC m=+956.850155620" Oct 01 07:21:45 crc kubenswrapper[4837]: I1001 07:21:45.093520 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-ktdtr" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.172594 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-grcxg"] Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.175624 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.178850 4837 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-hpl2z" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.188231 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-grcxg"] Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.259381 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h6r7\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-kube-api-access-9h6r7\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.259550 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.361183 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h6r7\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-kube-api-access-9h6r7\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.361353 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.389776 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.397885 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h6r7\" (UniqueName: \"kubernetes.io/projected/508d3925-1e4f-4775-b115-1e4aeccbf260-kube-api-access-9h6r7\") pod \"cert-manager-7d4cc89fcb-grcxg\" (UID: \"508d3925-1e4f-4775-b115-1e4aeccbf260\") " pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.517503 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" Oct 01 07:21:53 crc kubenswrapper[4837]: I1001 07:21:53.805499 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-grcxg"] Oct 01 07:21:53 crc kubenswrapper[4837]: W1001 07:21:53.812797 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod508d3925_1e4f_4775_b115_1e4aeccbf260.slice/crio-273e439e7fcb2690978183ae4cf164eb3cd9e307448bc16635123b6f9f3e2e92 WatchSource:0}: Error finding container 273e439e7fcb2690978183ae4cf164eb3cd9e307448bc16635123b6f9f3e2e92: Status 404 returned error can't find the container with id 273e439e7fcb2690978183ae4cf164eb3cd9e307448bc16635123b6f9f3e2e92 Oct 01 07:21:54 crc kubenswrapper[4837]: I1001 07:21:54.053980 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" event={"ID":"508d3925-1e4f-4775-b115-1e4aeccbf260","Type":"ContainerStarted","Data":"36435b370df4d4a42f2862ff15253ae383d4f3fd6dfb65e4d0ef0207615fa9c7"} Oct 01 07:21:54 crc kubenswrapper[4837]: I1001 07:21:54.054048 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" event={"ID":"508d3925-1e4f-4775-b115-1e4aeccbf260","Type":"ContainerStarted","Data":"273e439e7fcb2690978183ae4cf164eb3cd9e307448bc16635123b6f9f3e2e92"} Oct 01 07:21:54 crc kubenswrapper[4837]: I1001 07:21:54.084588 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-grcxg" podStartSLOduration=1.084561947 podStartE2EDuration="1.084561947s" podCreationTimestamp="2025-10-01 07:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:21:54.075818611 +0000 UTC m=+970.917426126" watchObservedRunningTime="2025-10-01 07:21:54.084561947 +0000 UTC m=+970.926169442" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.028666 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.030149 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.035147 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.040437 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9ct25" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.040484 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.043469 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.148062 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46mrp\" (UniqueName: \"kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp\") pod \"openstack-operator-index-6gqpd\" (UID: \"03471f22-8f6a-4663-a0a9-8a1b938807c3\") " pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.249956 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46mrp\" (UniqueName: \"kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp\") pod \"openstack-operator-index-6gqpd\" (UID: \"03471f22-8f6a-4663-a0a9-8a1b938807c3\") " pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.283460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46mrp\" (UniqueName: \"kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp\") pod \"openstack-operator-index-6gqpd\" (UID: \"03471f22-8f6a-4663-a0a9-8a1b938807c3\") " pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.359484 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:21:59 crc kubenswrapper[4837]: I1001 07:21:59.667518 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:22:00 crc kubenswrapper[4837]: I1001 07:22:00.098311 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6gqpd" event={"ID":"03471f22-8f6a-4663-a0a9-8a1b938807c3","Type":"ContainerStarted","Data":"84968a7a123439971b7db7e3b930e88978b6a61eb1a9b6ddaacb5149464972cb"} Oct 01 07:22:01 crc kubenswrapper[4837]: I1001 07:22:01.108360 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6gqpd" event={"ID":"03471f22-8f6a-4663-a0a9-8a1b938807c3","Type":"ContainerStarted","Data":"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b"} Oct 01 07:22:01 crc kubenswrapper[4837]: I1001 07:22:01.124999 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6gqpd" podStartSLOduration=2.23623729 podStartE2EDuration="3.124974955s" podCreationTimestamp="2025-10-01 07:21:58 +0000 UTC" firstStartedPulling="2025-10-01 07:21:59.67833536 +0000 UTC m=+976.519942825" lastFinishedPulling="2025-10-01 07:22:00.567072995 +0000 UTC m=+977.408680490" observedRunningTime="2025-10-01 07:22:01.123285263 +0000 UTC m=+977.964892748" watchObservedRunningTime="2025-10-01 07:22:01.124974955 +0000 UTC m=+977.966582450" Oct 01 07:22:02 crc kubenswrapper[4837]: I1001 07:22:02.373450 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:22:02 crc kubenswrapper[4837]: I1001 07:22:02.980680 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wvb74"] Oct 01 07:22:02 crc kubenswrapper[4837]: I1001 07:22:02.982893 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:02 crc kubenswrapper[4837]: I1001 07:22:02.988115 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wvb74"] Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.108921 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4xvr\" (UniqueName: \"kubernetes.io/projected/0dcc771f-a525-479f-b679-0783f280bb7a-kube-api-access-n4xvr\") pod \"openstack-operator-index-wvb74\" (UID: \"0dcc771f-a525-479f-b679-0783f280bb7a\") " pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.123208 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-6gqpd" podUID="03471f22-8f6a-4663-a0a9-8a1b938807c3" containerName="registry-server" containerID="cri-o://edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b" gracePeriod=2 Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.210199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4xvr\" (UniqueName: \"kubernetes.io/projected/0dcc771f-a525-479f-b679-0783f280bb7a-kube-api-access-n4xvr\") pod \"openstack-operator-index-wvb74\" (UID: \"0dcc771f-a525-479f-b679-0783f280bb7a\") " pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.255370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4xvr\" (UniqueName: \"kubernetes.io/projected/0dcc771f-a525-479f-b679-0783f280bb7a-kube-api-access-n4xvr\") pod \"openstack-operator-index-wvb74\" (UID: \"0dcc771f-a525-479f-b679-0783f280bb7a\") " pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.319796 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.629512 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wvb74"] Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.633426 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:22:03 crc kubenswrapper[4837]: W1001 07:22:03.633927 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dcc771f_a525_479f_b679_0783f280bb7a.slice/crio-b2071efc0c1fbefab86b1d39b44aac3e804658236ed87c55a76ae5b6e8a57518 WatchSource:0}: Error finding container b2071efc0c1fbefab86b1d39b44aac3e804658236ed87c55a76ae5b6e8a57518: Status 404 returned error can't find the container with id b2071efc0c1fbefab86b1d39b44aac3e804658236ed87c55a76ae5b6e8a57518 Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.719347 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46mrp\" (UniqueName: \"kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp\") pod \"03471f22-8f6a-4663-a0a9-8a1b938807c3\" (UID: \"03471f22-8f6a-4663-a0a9-8a1b938807c3\") " Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.724640 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp" (OuterVolumeSpecName: "kube-api-access-46mrp") pod "03471f22-8f6a-4663-a0a9-8a1b938807c3" (UID: "03471f22-8f6a-4663-a0a9-8a1b938807c3"). InnerVolumeSpecName "kube-api-access-46mrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:22:03 crc kubenswrapper[4837]: I1001 07:22:03.821370 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46mrp\" (UniqueName: \"kubernetes.io/projected/03471f22-8f6a-4663-a0a9-8a1b938807c3-kube-api-access-46mrp\") on node \"crc\" DevicePath \"\"" Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.136410 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wvb74" event={"ID":"0dcc771f-a525-479f-b679-0783f280bb7a","Type":"ContainerStarted","Data":"b2071efc0c1fbefab86b1d39b44aac3e804658236ed87c55a76ae5b6e8a57518"} Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.139413 4837 generic.go:334] "Generic (PLEG): container finished" podID="03471f22-8f6a-4663-a0a9-8a1b938807c3" containerID="edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b" exitCode=0 Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.139489 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6gqpd" event={"ID":"03471f22-8f6a-4663-a0a9-8a1b938807c3","Type":"ContainerDied","Data":"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b"} Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.139514 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6gqpd" Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.139541 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6gqpd" event={"ID":"03471f22-8f6a-4663-a0a9-8a1b938807c3","Type":"ContainerDied","Data":"84968a7a123439971b7db7e3b930e88978b6a61eb1a9b6ddaacb5149464972cb"} Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.139581 4837 scope.go:117] "RemoveContainer" containerID="edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b" Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.176608 4837 scope.go:117] "RemoveContainer" containerID="edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b" Oct 01 07:22:04 crc kubenswrapper[4837]: E1001 07:22:04.177892 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b\": container with ID starting with edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b not found: ID does not exist" containerID="edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b" Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.177972 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b"} err="failed to get container status \"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b\": rpc error: code = NotFound desc = could not find container \"edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b\": container with ID starting with edab5c007153c3e804b623320ef23537ec035a51ed531847e71f445c19ed8e2b not found: ID does not exist" Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.181907 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:22:04 crc kubenswrapper[4837]: I1001 07:22:04.194674 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-6gqpd"] Oct 01 07:22:05 crc kubenswrapper[4837]: I1001 07:22:05.150430 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wvb74" event={"ID":"0dcc771f-a525-479f-b679-0783f280bb7a","Type":"ContainerStarted","Data":"fc500eac2353edb7266a0e91fc575c8f2316beafe5fbcbe7fddc60e588883a4e"} Oct 01 07:22:05 crc kubenswrapper[4837]: I1001 07:22:05.180109 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wvb74" podStartSLOduration=2.713766054 podStartE2EDuration="3.180083795s" podCreationTimestamp="2025-10-01 07:22:02 +0000 UTC" firstStartedPulling="2025-10-01 07:22:03.639974835 +0000 UTC m=+980.481582320" lastFinishedPulling="2025-10-01 07:22:04.106292576 +0000 UTC m=+980.947900061" observedRunningTime="2025-10-01 07:22:05.173229055 +0000 UTC m=+982.014836560" watchObservedRunningTime="2025-10-01 07:22:05.180083795 +0000 UTC m=+982.021691290" Oct 01 07:22:05 crc kubenswrapper[4837]: I1001 07:22:05.825983 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03471f22-8f6a-4663-a0a9-8a1b938807c3" path="/var/lib/kubelet/pods/03471f22-8f6a-4663-a0a9-8a1b938807c3/volumes" Oct 01 07:22:13 crc kubenswrapper[4837]: I1001 07:22:13.320880 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:13 crc kubenswrapper[4837]: I1001 07:22:13.321729 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:13 crc kubenswrapper[4837]: I1001 07:22:13.369470 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:14 crc kubenswrapper[4837]: I1001 07:22:14.259540 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-wvb74" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.051328 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg"] Oct 01 07:22:16 crc kubenswrapper[4837]: E1001 07:22:16.052143 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03471f22-8f6a-4663-a0a9-8a1b938807c3" containerName="registry-server" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.052190 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="03471f22-8f6a-4663-a0a9-8a1b938807c3" containerName="registry-server" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.052476 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="03471f22-8f6a-4663-a0a9-8a1b938807c3" containerName="registry-server" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.054474 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.060196 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4b4ll" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.070823 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg"] Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.201141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.201249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hvq6\" (UniqueName: \"kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.201388 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.303054 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.303184 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.303259 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hvq6\" (UniqueName: \"kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.303758 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.304646 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.347317 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hvq6\" (UniqueName: \"kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6\") pod \"edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.387765 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:16 crc kubenswrapper[4837]: I1001 07:22:16.917980 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg"] Oct 01 07:22:17 crc kubenswrapper[4837]: I1001 07:22:17.252746 4837 generic.go:334] "Generic (PLEG): container finished" podID="deffc359-4840-433e-999e-fa7787e97ed2" containerID="98f6057627daefd727c643526b521a1fc86387e394bf62cbb1bc9aca463e600e" exitCode=0 Oct 01 07:22:17 crc kubenswrapper[4837]: I1001 07:22:17.252836 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" event={"ID":"deffc359-4840-433e-999e-fa7787e97ed2","Type":"ContainerDied","Data":"98f6057627daefd727c643526b521a1fc86387e394bf62cbb1bc9aca463e600e"} Oct 01 07:22:17 crc kubenswrapper[4837]: I1001 07:22:17.252897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" event={"ID":"deffc359-4840-433e-999e-fa7787e97ed2","Type":"ContainerStarted","Data":"5e2b4e5b69104d5204bbedd7d924053df91ba7f43396c99f4b097bda32e555d3"} Oct 01 07:22:18 crc kubenswrapper[4837]: I1001 07:22:18.263935 4837 generic.go:334] "Generic (PLEG): container finished" podID="deffc359-4840-433e-999e-fa7787e97ed2" containerID="314cf4409b2bb57207c89b627c30d72ee138e85721e7438add70d21e87a39882" exitCode=0 Oct 01 07:22:18 crc kubenswrapper[4837]: I1001 07:22:18.263985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" event={"ID":"deffc359-4840-433e-999e-fa7787e97ed2","Type":"ContainerDied","Data":"314cf4409b2bb57207c89b627c30d72ee138e85721e7438add70d21e87a39882"} Oct 01 07:22:19 crc kubenswrapper[4837]: I1001 07:22:19.278101 4837 generic.go:334] "Generic (PLEG): container finished" podID="deffc359-4840-433e-999e-fa7787e97ed2" containerID="b9c156a2125613fbbda78d484a59c634d07e8ea8ea15471bbeeeed23b00ff039" exitCode=0 Oct 01 07:22:19 crc kubenswrapper[4837]: I1001 07:22:19.278232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" event={"ID":"deffc359-4840-433e-999e-fa7787e97ed2","Type":"ContainerDied","Data":"b9c156a2125613fbbda78d484a59c634d07e8ea8ea15471bbeeeed23b00ff039"} Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.623569 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.677374 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util\") pod \"deffc359-4840-433e-999e-fa7787e97ed2\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.677551 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hvq6\" (UniqueName: \"kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6\") pod \"deffc359-4840-433e-999e-fa7787e97ed2\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.677599 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle\") pod \"deffc359-4840-433e-999e-fa7787e97ed2\" (UID: \"deffc359-4840-433e-999e-fa7787e97ed2\") " Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.678669 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle" (OuterVolumeSpecName: "bundle") pod "deffc359-4840-433e-999e-fa7787e97ed2" (UID: "deffc359-4840-433e-999e-fa7787e97ed2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.689266 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6" (OuterVolumeSpecName: "kube-api-access-8hvq6") pod "deffc359-4840-433e-999e-fa7787e97ed2" (UID: "deffc359-4840-433e-999e-fa7787e97ed2"). InnerVolumeSpecName "kube-api-access-8hvq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.710154 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util" (OuterVolumeSpecName: "util") pod "deffc359-4840-433e-999e-fa7787e97ed2" (UID: "deffc359-4840-433e-999e-fa7787e97ed2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.780134 4837 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.780212 4837 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/deffc359-4840-433e-999e-fa7787e97ed2-util\") on node \"crc\" DevicePath \"\"" Oct 01 07:22:20 crc kubenswrapper[4837]: I1001 07:22:20.780241 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hvq6\" (UniqueName: \"kubernetes.io/projected/deffc359-4840-433e-999e-fa7787e97ed2-kube-api-access-8hvq6\") on node \"crc\" DevicePath \"\"" Oct 01 07:22:21 crc kubenswrapper[4837]: I1001 07:22:21.299838 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" event={"ID":"deffc359-4840-433e-999e-fa7787e97ed2","Type":"ContainerDied","Data":"5e2b4e5b69104d5204bbedd7d924053df91ba7f43396c99f4b097bda32e555d3"} Oct 01 07:22:21 crc kubenswrapper[4837]: I1001 07:22:21.299882 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e2b4e5b69104d5204bbedd7d924053df91ba7f43396c99f4b097bda32e555d3" Oct 01 07:22:21 crc kubenswrapper[4837]: I1001 07:22:21.299910 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.745327 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br"] Oct 01 07:22:28 crc kubenswrapper[4837]: E1001 07:22:28.746326 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="util" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.746349 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="util" Oct 01 07:22:28 crc kubenswrapper[4837]: E1001 07:22:28.746364 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="pull" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.746376 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="pull" Oct 01 07:22:28 crc kubenswrapper[4837]: E1001 07:22:28.746395 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="extract" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.746408 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="extract" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.746655 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="deffc359-4840-433e-999e-fa7787e97ed2" containerName="extract" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.747962 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.750875 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-bscvs" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.784092 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br"] Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.890216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hqm5\" (UniqueName: \"kubernetes.io/projected/f53920b8-af09-4740-a121-b4b2207bba70-kube-api-access-9hqm5\") pod \"openstack-operator-controller-operator-8d79f897-jm6br\" (UID: \"f53920b8-af09-4740-a121-b4b2207bba70\") " pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:28 crc kubenswrapper[4837]: I1001 07:22:28.991316 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hqm5\" (UniqueName: \"kubernetes.io/projected/f53920b8-af09-4740-a121-b4b2207bba70-kube-api-access-9hqm5\") pod \"openstack-operator-controller-operator-8d79f897-jm6br\" (UID: \"f53920b8-af09-4740-a121-b4b2207bba70\") " pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:29 crc kubenswrapper[4837]: I1001 07:22:29.028241 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hqm5\" (UniqueName: \"kubernetes.io/projected/f53920b8-af09-4740-a121-b4b2207bba70-kube-api-access-9hqm5\") pod \"openstack-operator-controller-operator-8d79f897-jm6br\" (UID: \"f53920b8-af09-4740-a121-b4b2207bba70\") " pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:29 crc kubenswrapper[4837]: I1001 07:22:29.071004 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:29 crc kubenswrapper[4837]: I1001 07:22:29.356138 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br"] Oct 01 07:22:30 crc kubenswrapper[4837]: I1001 07:22:30.378206 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" event={"ID":"f53920b8-af09-4740-a121-b4b2207bba70","Type":"ContainerStarted","Data":"ac5876d774b87300633931d9aab776cff1eec4b8c1718c5354939862c3e059a5"} Oct 01 07:22:34 crc kubenswrapper[4837]: I1001 07:22:34.424100 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" event={"ID":"f53920b8-af09-4740-a121-b4b2207bba70","Type":"ContainerStarted","Data":"64c210032d2bd11616b49e1870e50dc13ac15a9bb1bbd1724d0835c0cb45414f"} Oct 01 07:22:36 crc kubenswrapper[4837]: I1001 07:22:36.440872 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" event={"ID":"f53920b8-af09-4740-a121-b4b2207bba70","Type":"ContainerStarted","Data":"855152e4bf1af081f8f0d1d4307a2406c81f2cde827993866bc2b4b18f5fc689"} Oct 01 07:22:36 crc kubenswrapper[4837]: I1001 07:22:36.441422 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:36 crc kubenswrapper[4837]: I1001 07:22:36.496486 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" podStartSLOduration=2.115614272 podStartE2EDuration="8.496456971s" podCreationTimestamp="2025-10-01 07:22:28 +0000 UTC" firstStartedPulling="2025-10-01 07:22:29.369014228 +0000 UTC m=+1006.210621683" lastFinishedPulling="2025-10-01 07:22:35.749856917 +0000 UTC m=+1012.591464382" observedRunningTime="2025-10-01 07:22:36.49316358 +0000 UTC m=+1013.334771075" watchObservedRunningTime="2025-10-01 07:22:36.496456971 +0000 UTC m=+1013.338064466" Oct 01 07:22:39 crc kubenswrapper[4837]: I1001 07:22:39.075057 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8d79f897-jm6br" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.186354 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.188293 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.191593 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-z4pwz" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.201146 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.207715 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.207853 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.214243 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-n5qlw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.256761 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.260485 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.261470 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.262986 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-flpfq" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.280376 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.281713 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.283320 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-n5rv7" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.294304 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.296207 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.301221 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.309935 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rcvbw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.323683 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.328480 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.334766 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.335722 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.339031 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8zrmm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.344709 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.345651 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.350991 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-m767z" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.353584 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.356577 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.360209 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.363610 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-48gfk" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.379753 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.380647 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdn2h\" (UniqueName: \"kubernetes.io/projected/06ebbbd0-256b-4637-b6e1-ba5adab9bd40-kube-api-access-fdn2h\") pod \"cinder-operator-controller-manager-859cd486d-lm9xd\" (UID: \"06ebbbd0-256b-4637-b6e1-ba5adab9bd40\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.380733 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldn95\" (UniqueName: \"kubernetes.io/projected/750df3b7-ef90-4f83-99f9-ebb54a6a159b-kube-api-access-ldn95\") pod \"glance-operator-controller-manager-8bc4775b5-wnqvm\" (UID: \"750df3b7-ef90-4f83-99f9-ebb54a6a159b\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.380762 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzwq7\" (UniqueName: \"kubernetes.io/projected/ebdcaccc-170c-40ff-9b06-57a1188cf20b-kube-api-access-vzwq7\") pod \"barbican-operator-controller-manager-f7f98cb69-rqr6f\" (UID: \"ebdcaccc-170c-40ff-9b06-57a1188cf20b\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.380843 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swblq\" (UniqueName: \"kubernetes.io/projected/e5e81392-cf50-43ab-b86a-e8881d5043f7-kube-api-access-swblq\") pod \"designate-operator-controller-manager-77fb7bcf5b-lq7sr\" (UID: \"e5e81392-cf50-43ab-b86a-e8881d5043f7\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.381036 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.383529 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-f654s" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.388267 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.396758 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.397767 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.402130 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-n6fb8" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.410408 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.416424 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.417618 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.419061 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pz2vj" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.424772 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.433942 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.438737 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.449889 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.450823 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.452756 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.455377 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-69kdq" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.463923 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.464886 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.468029 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-m5m9h" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.468097 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.479209 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.482757 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.483753 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484080 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzwq7\" (UniqueName: \"kubernetes.io/projected/ebdcaccc-170c-40ff-9b06-57a1188cf20b-kube-api-access-vzwq7\") pod \"barbican-operator-controller-manager-f7f98cb69-rqr6f\" (UID: \"ebdcaccc-170c-40ff-9b06-57a1188cf20b\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484154 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8g8r\" (UniqueName: \"kubernetes.io/projected/0cd7b250-33bc-45fc-be0d-e4d6d28e450f-kube-api-access-h8g8r\") pod \"horizon-operator-controller-manager-679b4759bb-rz6jg\" (UID: \"0cd7b250-33bc-45fc-be0d-e4d6d28e450f\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484180 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swblq\" (UniqueName: \"kubernetes.io/projected/e5e81392-cf50-43ab-b86a-e8881d5043f7-kube-api-access-swblq\") pod \"designate-operator-controller-manager-77fb7bcf5b-lq7sr\" (UID: \"e5e81392-cf50-43ab-b86a-e8881d5043f7\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484199 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478t8\" (UniqueName: \"kubernetes.io/projected/426cfad6-045e-48e1-82e4-ce9b9891dd0a-kube-api-access-478t8\") pod \"heat-operator-controller-manager-5b4fc86755-95dns\" (UID: \"426cfad6-045e-48e1-82e4-ce9b9891dd0a\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484224 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lg2d\" (UniqueName: \"kubernetes.io/projected/5d9a421b-7104-462a-8a01-8bc62522a543-kube-api-access-4lg2d\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484246 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdn2h\" (UniqueName: \"kubernetes.io/projected/06ebbbd0-256b-4637-b6e1-ba5adab9bd40-kube-api-access-fdn2h\") pod \"cinder-operator-controller-manager-859cd486d-lm9xd\" (UID: \"06ebbbd0-256b-4637-b6e1-ba5adab9bd40\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484279 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9mlc\" (UniqueName: \"kubernetes.io/projected/f987881b-66b4-49fe-a171-15813e9225f5-kube-api-access-d9mlc\") pod \"ironic-operator-controller-manager-6f589bc7f7-wf96x\" (UID: \"f987881b-66b4-49fe-a171-15813e9225f5\") " pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484295 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.484317 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldn95\" (UniqueName: \"kubernetes.io/projected/750df3b7-ef90-4f83-99f9-ebb54a6a159b-kube-api-access-ldn95\") pod \"glance-operator-controller-manager-8bc4775b5-wnqvm\" (UID: \"750df3b7-ef90-4f83-99f9-ebb54a6a159b\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.487565 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-c6s6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.498097 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.507771 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzwq7\" (UniqueName: \"kubernetes.io/projected/ebdcaccc-170c-40ff-9b06-57a1188cf20b-kube-api-access-vzwq7\") pod \"barbican-operator-controller-manager-f7f98cb69-rqr6f\" (UID: \"ebdcaccc-170c-40ff-9b06-57a1188cf20b\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.508005 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.525265 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.527336 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nvs79" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.530206 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldn95\" (UniqueName: \"kubernetes.io/projected/750df3b7-ef90-4f83-99f9-ebb54a6a159b-kube-api-access-ldn95\") pod \"glance-operator-controller-manager-8bc4775b5-wnqvm\" (UID: \"750df3b7-ef90-4f83-99f9-ebb54a6a159b\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.530238 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdn2h\" (UniqueName: \"kubernetes.io/projected/06ebbbd0-256b-4637-b6e1-ba5adab9bd40-kube-api-access-fdn2h\") pod \"cinder-operator-controller-manager-859cd486d-lm9xd\" (UID: \"06ebbbd0-256b-4637-b6e1-ba5adab9bd40\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.536187 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.557113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swblq\" (UniqueName: \"kubernetes.io/projected/e5e81392-cf50-43ab-b86a-e8881d5043f7-kube-api-access-swblq\") pod \"designate-operator-controller-manager-77fb7bcf5b-lq7sr\" (UID: \"e5e81392-cf50-43ab-b86a-e8881d5043f7\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.562967 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.576950 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.593674 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.593747 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-822wp"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.594560 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.594816 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.600759 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.601157 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mlzdm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.601324 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-f2hz7" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.604370 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcrgc\" (UniqueName: \"kubernetes.io/projected/fef6cdf1-8dbe-47e9-8d8a-143710d954ec-kube-api-access-wcrgc\") pod \"neutron-operator-controller-manager-6b96467f46-kgf8r\" (UID: \"fef6cdf1-8dbe-47e9-8d8a-143710d954ec\") " pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.604454 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9mlc\" (UniqueName: \"kubernetes.io/projected/f987881b-66b4-49fe-a171-15813e9225f5-kube-api-access-d9mlc\") pod \"ironic-operator-controller-manager-6f589bc7f7-wf96x\" (UID: \"f987881b-66b4-49fe-a171-15813e9225f5\") " pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.604906 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.604975 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqkx\" (UniqueName: \"kubernetes.io/projected/b537c4f7-697f-4856-b7f9-98a55150cc9f-kube-api-access-kwqkx\") pod \"keystone-operator-controller-manager-59d7dc95cf-6hgks\" (UID: \"b537c4f7-697f-4856-b7f9-98a55150cc9f\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.605006 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q"] Oct 01 07:22:56 crc kubenswrapper[4837]: E1001 07:22:56.605633 4837 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 01 07:22:56 crc kubenswrapper[4837]: E1001 07:22:56.605819 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert podName:5d9a421b-7104-462a-8a01-8bc62522a543 nodeName:}" failed. No retries permitted until 2025-10-01 07:22:57.105799211 +0000 UTC m=+1033.947406666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert") pod "infra-operator-controller-manager-5c8fdc4d5c-tlhpd" (UID: "5d9a421b-7104-462a-8a01-8bc62522a543") : secret "infra-operator-webhook-server-cert" not found Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.606032 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609045 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609476 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll4xb\" (UniqueName: \"kubernetes.io/projected/2ed39459-5a70-4c13-aacd-966aae697780-kube-api-access-ll4xb\") pod \"octavia-operator-controller-manager-6fb7d6b8bf-2xtdw\" (UID: \"2ed39459-5a70-4c13-aacd-966aae697780\") " pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609523 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkgk2\" (UniqueName: \"kubernetes.io/projected/3efcdf55-46eb-446e-933a-e086f3a67aab-kube-api-access-mkgk2\") pod \"manila-operator-controller-manager-b7cf8cb5f-l4pdl\" (UID: \"3efcdf55-46eb-446e-933a-e086f3a67aab\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609568 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8g8r\" (UniqueName: \"kubernetes.io/projected/0cd7b250-33bc-45fc-be0d-e4d6d28e450f-kube-api-access-h8g8r\") pod \"horizon-operator-controller-manager-679b4759bb-rz6jg\" (UID: \"0cd7b250-33bc-45fc-be0d-e4d6d28e450f\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609596 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lcsq\" (UniqueName: \"kubernetes.io/projected/e6fb38c5-e91c-4229-9aa5-93729ffab045-kube-api-access-4lcsq\") pod \"mariadb-operator-controller-manager-67bf5bb885-mgm24\" (UID: \"e6fb38c5-e91c-4229-9aa5-93729ffab045\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609631 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478t8\" (UniqueName: \"kubernetes.io/projected/426cfad6-045e-48e1-82e4-ce9b9891dd0a-kube-api-access-478t8\") pod \"heat-operator-controller-manager-5b4fc86755-95dns\" (UID: \"426cfad6-045e-48e1-82e4-ce9b9891dd0a\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.609663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2pb8\" (UniqueName: \"kubernetes.io/projected/76e4d2ee-d38f-433c-945c-29f36594a338-kube-api-access-d2pb8\") pod \"nova-operator-controller-manager-79f9fc9fd8-p978w\" (UID: \"76e4d2ee-d38f-433c-945c-29f36594a338\") " pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.610428 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lg2d\" (UniqueName: \"kubernetes.io/projected/5d9a421b-7104-462a-8a01-8bc62522a543-kube-api-access-4lg2d\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.636416 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9mlc\" (UniqueName: \"kubernetes.io/projected/f987881b-66b4-49fe-a171-15813e9225f5-kube-api-access-d9mlc\") pod \"ironic-operator-controller-manager-6f589bc7f7-wf96x\" (UID: \"f987881b-66b4-49fe-a171-15813e9225f5\") " pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.636853 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-822wp"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.636963 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8g8r\" (UniqueName: \"kubernetes.io/projected/0cd7b250-33bc-45fc-be0d-e4d6d28e450f-kube-api-access-h8g8r\") pod \"horizon-operator-controller-manager-679b4759bb-rz6jg\" (UID: \"0cd7b250-33bc-45fc-be0d-e4d6d28e450f\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.638746 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478t8\" (UniqueName: \"kubernetes.io/projected/426cfad6-045e-48e1-82e4-ce9b9891dd0a-kube-api-access-478t8\") pod \"heat-operator-controller-manager-5b4fc86755-95dns\" (UID: \"426cfad6-045e-48e1-82e4-ce9b9891dd0a\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.645458 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.647114 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.649876 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-2xtk2" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.652145 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.657128 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lg2d\" (UniqueName: \"kubernetes.io/projected/5d9a421b-7104-462a-8a01-8bc62522a543-kube-api-access-4lg2d\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.663740 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.690088 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.691386 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.694223 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.694264 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-q9fft" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.697006 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712202 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcrgc\" (UniqueName: \"kubernetes.io/projected/fef6cdf1-8dbe-47e9-8d8a-143710d954ec-kube-api-access-wcrgc\") pod \"neutron-operator-controller-manager-6b96467f46-kgf8r\" (UID: \"fef6cdf1-8dbe-47e9-8d8a-143710d954ec\") " pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712291 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqkx\" (UniqueName: \"kubernetes.io/projected/b537c4f7-697f-4856-b7f9-98a55150cc9f-kube-api-access-kwqkx\") pod \"keystone-operator-controller-manager-59d7dc95cf-6hgks\" (UID: \"b537c4f7-697f-4856-b7f9-98a55150cc9f\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712347 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll4xb\" (UniqueName: \"kubernetes.io/projected/2ed39459-5a70-4c13-aacd-966aae697780-kube-api-access-ll4xb\") pod \"octavia-operator-controller-manager-6fb7d6b8bf-2xtdw\" (UID: \"2ed39459-5a70-4c13-aacd-966aae697780\") " pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712376 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb4hz\" (UniqueName: \"kubernetes.io/projected/d4fc2a84-086a-4644-b354-2ef743bf1e7f-kube-api-access-rb4hz\") pod \"placement-operator-controller-manager-598c4c8547-822wp\" (UID: \"d4fc2a84-086a-4644-b354-2ef743bf1e7f\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712398 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkgk2\" (UniqueName: \"kubernetes.io/projected/3efcdf55-46eb-446e-933a-e086f3a67aab-kube-api-access-mkgk2\") pod \"manila-operator-controller-manager-b7cf8cb5f-l4pdl\" (UID: \"3efcdf55-46eb-446e-933a-e086f3a67aab\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712421 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5mvp\" (UniqueName: \"kubernetes.io/projected/969f8b13-b131-4fad-84ff-6b092fcc43e9-kube-api-access-g5mvp\") pod \"ovn-operator-controller-manager-84c745747f-6bm6v\" (UID: \"969f8b13-b131-4fad-84ff-6b092fcc43e9\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712462 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lcsq\" (UniqueName: \"kubernetes.io/projected/e6fb38c5-e91c-4229-9aa5-93729ffab045-kube-api-access-4lcsq\") pod \"mariadb-operator-controller-manager-67bf5bb885-mgm24\" (UID: \"e6fb38c5-e91c-4229-9aa5-93729ffab045\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712501 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2pb8\" (UniqueName: \"kubernetes.io/projected/76e4d2ee-d38f-433c-945c-29f36594a338-kube-api-access-d2pb8\") pod \"nova-operator-controller-manager-79f9fc9fd8-p978w\" (UID: \"76e4d2ee-d38f-433c-945c-29f36594a338\") " pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.712536 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lbf5\" (UniqueName: \"kubernetes.io/projected/648417c5-8d92-4426-81db-052bcb20cbc6-kube-api-access-7lbf5\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.721904 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.723211 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.728900 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-g4j5q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.734819 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqkx\" (UniqueName: \"kubernetes.io/projected/b537c4f7-697f-4856-b7f9-98a55150cc9f-kube-api-access-kwqkx\") pod \"keystone-operator-controller-manager-59d7dc95cf-6hgks\" (UID: \"b537c4f7-697f-4856-b7f9-98a55150cc9f\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.735213 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.736205 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcrgc\" (UniqueName: \"kubernetes.io/projected/fef6cdf1-8dbe-47e9-8d8a-143710d954ec-kube-api-access-wcrgc\") pod \"neutron-operator-controller-manager-6b96467f46-kgf8r\" (UID: \"fef6cdf1-8dbe-47e9-8d8a-143710d954ec\") " pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.736203 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lcsq\" (UniqueName: \"kubernetes.io/projected/e6fb38c5-e91c-4229-9aa5-93729ffab045-kube-api-access-4lcsq\") pod \"mariadb-operator-controller-manager-67bf5bb885-mgm24\" (UID: \"e6fb38c5-e91c-4229-9aa5-93729ffab045\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.736936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll4xb\" (UniqueName: \"kubernetes.io/projected/2ed39459-5a70-4c13-aacd-966aae697780-kube-api-access-ll4xb\") pod \"octavia-operator-controller-manager-6fb7d6b8bf-2xtdw\" (UID: \"2ed39459-5a70-4c13-aacd-966aae697780\") " pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.749232 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.750506 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2pb8\" (UniqueName: \"kubernetes.io/projected/76e4d2ee-d38f-433c-945c-29f36594a338-kube-api-access-d2pb8\") pod \"nova-operator-controller-manager-79f9fc9fd8-p978w\" (UID: \"76e4d2ee-d38f-433c-945c-29f36594a338\") " pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.755325 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkgk2\" (UniqueName: \"kubernetes.io/projected/3efcdf55-46eb-446e-933a-e086f3a67aab-kube-api-access-mkgk2\") pod \"manila-operator-controller-manager-b7cf8cb5f-l4pdl\" (UID: \"3efcdf55-46eb-446e-933a-e086f3a67aab\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.780681 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.794317 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.795675 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.798339 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.798963 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.802586 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jq6c5" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.804578 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.818900 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b6l7\" (UniqueName: \"kubernetes.io/projected/414c531b-a0c8-4cb4-a997-77706c2f8e7c-kube-api-access-2b6l7\") pod \"watcher-operator-controller-manager-75756dd4d9-vx2m2\" (UID: \"414c531b-a0c8-4cb4-a997-77706c2f8e7c\") " pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.818961 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb4hz\" (UniqueName: \"kubernetes.io/projected/d4fc2a84-086a-4644-b354-2ef743bf1e7f-kube-api-access-rb4hz\") pod \"placement-operator-controller-manager-598c4c8547-822wp\" (UID: \"d4fc2a84-086a-4644-b354-2ef743bf1e7f\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.818989 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.819007 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5mvp\" (UniqueName: \"kubernetes.io/projected/969f8b13-b131-4fad-84ff-6b092fcc43e9-kube-api-access-g5mvp\") pod \"ovn-operator-controller-manager-84c745747f-6bm6v\" (UID: \"969f8b13-b131-4fad-84ff-6b092fcc43e9\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.819028 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk6cw\" (UniqueName: \"kubernetes.io/projected/70c21bc2-f519-4551-bf8d-669df6064834-kube-api-access-rk6cw\") pod \"swift-operator-controller-manager-657c6b68c7-fddfz\" (UID: \"70c21bc2-f519-4551-bf8d-669df6064834\") " pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.819044 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zngsh\" (UniqueName: \"kubernetes.io/projected/2297d6d0-957e-4d04-9194-d9d7c120c3f0-kube-api-access-zngsh\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmqwh\" (UID: \"2297d6d0-957e-4d04-9194-d9d7c120c3f0\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.819083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lbf5\" (UniqueName: \"kubernetes.io/projected/648417c5-8d92-4426-81db-052bcb20cbc6-kube-api-access-7lbf5\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: E1001 07:22:56.820079 4837 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 07:22:56 crc kubenswrapper[4837]: E1001 07:22:56.820197 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert podName:648417c5-8d92-4426-81db-052bcb20cbc6 nodeName:}" failed. No retries permitted until 2025-10-01 07:22:57.320182121 +0000 UTC m=+1034.161789576 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert") pod "openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" (UID: "648417c5-8d92-4426-81db-052bcb20cbc6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.839362 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lbf5\" (UniqueName: \"kubernetes.io/projected/648417c5-8d92-4426-81db-052bcb20cbc6-kube-api-access-7lbf5\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.844471 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5mvp\" (UniqueName: \"kubernetes.io/projected/969f8b13-b131-4fad-84ff-6b092fcc43e9-kube-api-access-g5mvp\") pod \"ovn-operator-controller-manager-84c745747f-6bm6v\" (UID: \"969f8b13-b131-4fad-84ff-6b092fcc43e9\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.845798 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb4hz\" (UniqueName: \"kubernetes.io/projected/d4fc2a84-086a-4644-b354-2ef743bf1e7f-kube-api-access-rb4hz\") pod \"placement-operator-controller-manager-598c4c8547-822wp\" (UID: \"d4fc2a84-086a-4644-b354-2ef743bf1e7f\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.893025 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.916610 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.917622 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.921248 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.921559 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4dp75" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.922841 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk6cw\" (UniqueName: \"kubernetes.io/projected/70c21bc2-f519-4551-bf8d-669df6064834-kube-api-access-rk6cw\") pod \"swift-operator-controller-manager-657c6b68c7-fddfz\" (UID: \"70c21bc2-f519-4551-bf8d-669df6064834\") " pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.922885 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zngsh\" (UniqueName: \"kubernetes.io/projected/2297d6d0-957e-4d04-9194-d9d7c120c3f0-kube-api-access-zngsh\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmqwh\" (UID: \"2297d6d0-957e-4d04-9194-d9d7c120c3f0\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.922949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n655x\" (UniqueName: \"kubernetes.io/projected/0a3d4dda-cb7c-48ae-95ee-f6eebe98745e-kube-api-access-n655x\") pod \"test-operator-controller-manager-6bb97fcf96-vcm95\" (UID: \"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e\") " pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.923016 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b6l7\" (UniqueName: \"kubernetes.io/projected/414c531b-a0c8-4cb4-a997-77706c2f8e7c-kube-api-access-2b6l7\") pod \"watcher-operator-controller-manager-75756dd4d9-vx2m2\" (UID: \"414c531b-a0c8-4cb4-a997-77706c2f8e7c\") " pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.930458 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.931340 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.936434 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.940443 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk6cw\" (UniqueName: \"kubernetes.io/projected/70c21bc2-f519-4551-bf8d-669df6064834-kube-api-access-rk6cw\") pod \"swift-operator-controller-manager-657c6b68c7-fddfz\" (UID: \"70c21bc2-f519-4551-bf8d-669df6064834\") " pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.947040 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zngsh\" (UniqueName: \"kubernetes.io/projected/2297d6d0-957e-4d04-9194-d9d7c120c3f0-kube-api-access-zngsh\") pod \"telemetry-operator-controller-manager-cb66d6b59-cmqwh\" (UID: \"2297d6d0-957e-4d04-9194-d9d7c120c3f0\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.982936 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.983763 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.987667 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-ptjj5" Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.989752 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2"] Oct 01 07:22:56 crc kubenswrapper[4837]: I1001 07:22:56.998318 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b6l7\" (UniqueName: \"kubernetes.io/projected/414c531b-a0c8-4cb4-a997-77706c2f8e7c-kube-api-access-2b6l7\") pod \"watcher-operator-controller-manager-75756dd4d9-vx2m2\" (UID: \"414c531b-a0c8-4cb4-a997-77706c2f8e7c\") " pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.002959 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.019753 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.021077 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.023872 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxjmp\" (UniqueName: \"kubernetes.io/projected/7de7f612-a5bd-4f2c-a146-d5f49ba757af-kube-api-access-cxjmp\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.023905 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.023939 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n655x\" (UniqueName: \"kubernetes.io/projected/0a3d4dda-cb7c-48ae-95ee-f6eebe98745e-kube-api-access-n655x\") pod \"test-operator-controller-manager-6bb97fcf96-vcm95\" (UID: \"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e\") " pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.037326 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.041807 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n655x\" (UniqueName: \"kubernetes.io/projected/0a3d4dda-cb7c-48ae-95ee-f6eebe98745e-kube-api-access-n655x\") pod \"test-operator-controller-manager-6bb97fcf96-vcm95\" (UID: \"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e\") " pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.067152 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.128899 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.128976 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddr7h\" (UniqueName: \"kubernetes.io/projected/421ed6b2-06bf-421e-a871-ec953b32bc93-kube-api-access-ddr7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2\" (UID: \"421ed6b2-06bf-421e-a871-ec953b32bc93\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.129053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxjmp\" (UniqueName: \"kubernetes.io/projected/7de7f612-a5bd-4f2c-a146-d5f49ba757af-kube-api-access-cxjmp\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.129077 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: E1001 07:22:57.129200 4837 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 01 07:22:57 crc kubenswrapper[4837]: E1001 07:22:57.129252 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert podName:7de7f612-a5bd-4f2c-a146-d5f49ba757af nodeName:}" failed. No retries permitted until 2025-10-01 07:22:57.629238286 +0000 UTC m=+1034.470845741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert") pod "openstack-operator-controller-manager-8648d97544-8vjp6" (UID: "7de7f612-a5bd-4f2c-a146-d5f49ba757af") : secret "webhook-server-cert" not found Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.139130 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.155840 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d9a421b-7104-462a-8a01-8bc62522a543-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-tlhpd\" (UID: \"5d9a421b-7104-462a-8a01-8bc62522a543\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.176252 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxjmp\" (UniqueName: \"kubernetes.io/projected/7de7f612-a5bd-4f2c-a146-d5f49ba757af-kube-api-access-cxjmp\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.231796 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddr7h\" (UniqueName: \"kubernetes.io/projected/421ed6b2-06bf-421e-a871-ec953b32bc93-kube-api-access-ddr7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2\" (UID: \"421ed6b2-06bf-421e-a871-ec953b32bc93\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.256833 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddr7h\" (UniqueName: \"kubernetes.io/projected/421ed6b2-06bf-421e-a871-ec953b32bc93-kube-api-access-ddr7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2\" (UID: \"421ed6b2-06bf-421e-a871-ec953b32bc93\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.288491 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.334215 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.334940 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.380387 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/648417c5-8d92-4426-81db-052bcb20cbc6-cert\") pod \"openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q\" (UID: \"648417c5-8d92-4426-81db-052bcb20cbc6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.588912 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.637305 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.642092 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7de7f612-a5bd-4f2c-a146-d5f49ba757af-cert\") pod \"openstack-operator-controller-manager-8648d97544-8vjp6\" (UID: \"7de7f612-a5bd-4f2c-a146-d5f49ba757af\") " pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.808620 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f"] Oct 01 07:22:57 crc kubenswrapper[4837]: W1001 07:22:57.820670 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebdcaccc_170c_40ff_9b06_57a1188cf20b.slice/crio-5c1a312b2a6774adce5d3f7a10ef5d33662a1b83cf8073331fe2b0d197ba0b30 WatchSource:0}: Error finding container 5c1a312b2a6774adce5d3f7a10ef5d33662a1b83cf8073331fe2b0d197ba0b30: Status 404 returned error can't find the container with id 5c1a312b2a6774adce5d3f7a10ef5d33662a1b83cf8073331fe2b0d197ba0b30 Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.852387 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr"] Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.861079 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd"] Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.919538 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.945671 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg"] Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.954036 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm"] Oct 01 07:22:57 crc kubenswrapper[4837]: I1001 07:22:57.966428 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.473553 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.476358 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.486825 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.497307 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.519371 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.523413 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-822wp"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.536413 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.554080 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.554138 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.558402 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.562681 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.568335 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.572042 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.575352 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.578770 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.583192 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.591350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" event={"ID":"f987881b-66b4-49fe-a171-15813e9225f5","Type":"ContainerStarted","Data":"161e05c447ce8f8273edc2bd41c3e49a5d0638b74325790c2cd4a100762a69ba"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.596826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" event={"ID":"e5e81392-cf50-43ab-b86a-e8881d5043f7","Type":"ContainerStarted","Data":"fc5c9717bd57b90fde7204787814869ea3dc89df44172fdd9f6f60e7c0be398e"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.597489 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6"] Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.599144 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" event={"ID":"b537c4f7-697f-4856-b7f9-98a55150cc9f","Type":"ContainerStarted","Data":"6bd587f54b9d2fc4dcf42b52caa8db55cc4dd2f96bd07b1eb1e7aa6f91a0f9f7"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.602851 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" event={"ID":"750df3b7-ef90-4f83-99f9-ebb54a6a159b","Type":"ContainerStarted","Data":"e04a5b9e84aeb192cebfbacd177a18bebc68a30f48e68a49f2a839d1785d3108"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.604594 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" event={"ID":"0cd7b250-33bc-45fc-be0d-e4d6d28e450f","Type":"ContainerStarted","Data":"4bf860be1b3d376c8742a666fede4a937045c6e3c6d183046bc9e78dd9aa5461"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.607101 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" event={"ID":"76e4d2ee-d38f-433c-945c-29f36594a338","Type":"ContainerStarted","Data":"f60eeb4b001059c402a9ecb132f5dcc76d201fc91d2645d336a0ddb3dd947802"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.610633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" event={"ID":"2ed39459-5a70-4c13-aacd-966aae697780","Type":"ContainerStarted","Data":"b873d71e312c69a91f1e91ab571167633b3b23ad5c53862a17c463ea165b6f9f"} Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.610963 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" event={"ID":"ebdcaccc-170c-40ff-9b06-57a1188cf20b","Type":"ContainerStarted","Data":"5c1a312b2a6774adce5d3f7a10ef5d33662a1b83cf8073331fe2b0d197ba0b30"} Oct 01 07:22:58 crc kubenswrapper[4837]: W1001 07:22:58.611316 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d9a421b_7104_462a_8a01_8bc62522a543.slice/crio-5928dd1f896b68e5026b82bf3558fb9710effef276dd461bd0ebca98dfd5fa2c WatchSource:0}: Error finding container 5928dd1f896b68e5026b82bf3558fb9710effef276dd461bd0ebca98dfd5fa2c: Status 404 returned error can't find the container with id 5928dd1f896b68e5026b82bf3558fb9710effef276dd461bd0ebca98dfd5fa2c Oct 01 07:22:58 crc kubenswrapper[4837]: I1001 07:22:58.611885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" event={"ID":"06ebbbd0-256b-4637-b6e1-ba5adab9bd40","Type":"ContainerStarted","Data":"b6cf61f111158a45011f1fbf820a9055cb3be92d0e2d212d9d01c855b2b4c959"} Oct 01 07:22:58 crc kubenswrapper[4837]: W1001 07:22:58.612727 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod648417c5_8d92_4426_81db_052bcb20cbc6.slice/crio-8aff0e1a5e3003f172572f5ce8a7806ace066da7b3a444e62255e8b01fbe0bd5 WatchSource:0}: Error finding container 8aff0e1a5e3003f172572f5ce8a7806ace066da7b3a444e62255e8b01fbe0bd5: Status 404 returned error can't find the container with id 8aff0e1a5e3003f172572f5ce8a7806ace066da7b3a444e62255e8b01fbe0bd5 Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.619370 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-478t8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b4fc86755-95dns_openstack-operators(426cfad6-045e-48e1-82e4-ce9b9891dd0a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.653824 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2b6l7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75756dd4d9-vx2m2_openstack-operators(414c531b-a0c8-4cb4-a997-77706c2f8e7c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.666401 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zngsh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-cb66d6b59-cmqwh_openstack-operators(2297d6d0-957e-4d04-9194-d9d7c120c3f0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: W1001 07:22:58.676596 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod969f8b13_b131_4fad_84ff_6b092fcc43e9.slice/crio-4e44ca51b0f5a09dc52a126761706a50d3f8214d0dfe2158dbbbca2619065a07 WatchSource:0}: Error finding container 4e44ca51b0f5a09dc52a126761706a50d3f8214d0dfe2158dbbbca2619065a07: Status 404 returned error can't find the container with id 4e44ca51b0f5a09dc52a126761706a50d3f8214d0dfe2158dbbbca2619065a07 Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.735982 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g5mvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-84c745747f-6bm6v_openstack-operators(969f8b13-b131-4fad-84ff-6b092fcc43e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.736453 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rb4hz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-598c4c8547-822wp_openstack-operators(d4fc2a84-086a-4644-b354-2ef743bf1e7f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.736479 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mkgk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-b7cf8cb5f-l4pdl_openstack-operators(3efcdf55-46eb-446e-933a-e086f3a67aab): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.736533 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n655x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb97fcf96-vcm95_openstack-operators(0a3d4dda-cb7c-48ae-95ee-f6eebe98745e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.736773 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddr7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2_openstack-operators(421ed6b2-06bf-421e-a871-ec953b32bc93): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.739832 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" podUID="421ed6b2-06bf-421e-a871-ec953b32bc93" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.876115 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" podUID="426cfad6-045e-48e1-82e4-ce9b9891dd0a" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.887775 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" podUID="414c531b-a0c8-4cb4-a997-77706c2f8e7c" Oct 01 07:22:58 crc kubenswrapper[4837]: E1001 07:22:58.945104 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" podUID="2297d6d0-957e-4d04-9194-d9d7c120c3f0" Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.022873 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" podUID="d4fc2a84-086a-4644-b354-2ef743bf1e7f" Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.030143 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" podUID="3efcdf55-46eb-446e-933a-e086f3a67aab" Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.044095 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" podUID="0a3d4dda-cb7c-48ae-95ee-f6eebe98745e" Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.057948 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" podUID="969f8b13-b131-4fad-84ff-6b092fcc43e9" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.631628 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" event={"ID":"421ed6b2-06bf-421e-a871-ec953b32bc93","Type":"ContainerStarted","Data":"47065f93d1bece24bfcb1b1c47183bbc0e639bc376e75ca9df3b05b3ffb1ce10"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.634528 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" podUID="421ed6b2-06bf-421e-a871-ec953b32bc93" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.635764 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" event={"ID":"70c21bc2-f519-4551-bf8d-669df6064834","Type":"ContainerStarted","Data":"2a9697b6858611dfab6b8dc439d3845516c473446f6bf3a4932c025c42b011fd"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.638671 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" event={"ID":"d4fc2a84-086a-4644-b354-2ef743bf1e7f","Type":"ContainerStarted","Data":"2180f4e7494113ecbfdcd80eadaafa089c5c97a7f9ea2298945faf49c8f05d6b"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.638719 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" event={"ID":"d4fc2a84-086a-4644-b354-2ef743bf1e7f","Type":"ContainerStarted","Data":"013ac4194e7f274c85a65e71c32a6bcd5544b5f1d1e5cbb2fea42b8fedc19f8e"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.639446 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" podUID="d4fc2a84-086a-4644-b354-2ef743bf1e7f" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.642487 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" event={"ID":"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e","Type":"ContainerStarted","Data":"ff5dc12d6a0b19b7fd443de1653987b4321328505ae57cf50f0bdbcf18952909"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.642511 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" event={"ID":"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e","Type":"ContainerStarted","Data":"fc5e28bf40f728af0fab3f31fef6cb443f8441b7791ec4d5b4c83cabf2f4e3c2"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.663994 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" event={"ID":"3efcdf55-46eb-446e-933a-e086f3a67aab","Type":"ContainerStarted","Data":"c896ac8b8d27ea74d8015a49c155bb2b7467fde544da8dc8803c4e09f4f8fd56"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.664035 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" event={"ID":"3efcdf55-46eb-446e-933a-e086f3a67aab","Type":"ContainerStarted","Data":"d39aa28b87794ed4121bb762cc3103220df216ddaa86959d9ad267a6d6fffcae"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.664073 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" podUID="0a3d4dda-cb7c-48ae-95ee-f6eebe98745e" Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.667456 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" podUID="3efcdf55-46eb-446e-933a-e086f3a67aab" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.668426 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" event={"ID":"969f8b13-b131-4fad-84ff-6b092fcc43e9","Type":"ContainerStarted","Data":"e4748b0b9f397120bfcde57926558167095e1fe4624dce019b5d1fcead0a16c0"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.668455 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" event={"ID":"969f8b13-b131-4fad-84ff-6b092fcc43e9","Type":"ContainerStarted","Data":"4e44ca51b0f5a09dc52a126761706a50d3f8214d0dfe2158dbbbca2619065a07"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.674099 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" podUID="969f8b13-b131-4fad-84ff-6b092fcc43e9" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.679623 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" event={"ID":"648417c5-8d92-4426-81db-052bcb20cbc6","Type":"ContainerStarted","Data":"8aff0e1a5e3003f172572f5ce8a7806ace066da7b3a444e62255e8b01fbe0bd5"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.692626 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" event={"ID":"e6fb38c5-e91c-4229-9aa5-93729ffab045","Type":"ContainerStarted","Data":"70e1fed901c2d14bba072578d535f22f5f71f7babc05ede1afba612fa364430d"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.723565 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" event={"ID":"2297d6d0-957e-4d04-9194-d9d7c120c3f0","Type":"ContainerStarted","Data":"317d57e7a85ce1ff9c6c2b0d40c730fd290c8f6afe2848319d3299f06ef41a46"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.723633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" event={"ID":"2297d6d0-957e-4d04-9194-d9d7c120c3f0","Type":"ContainerStarted","Data":"1c45375a49a9065509b06b15e071c3653c28fab77d15baa10efcc4c2c2c5e906"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.725951 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" podUID="2297d6d0-957e-4d04-9194-d9d7c120c3f0" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.727419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" event={"ID":"414c531b-a0c8-4cb4-a997-77706c2f8e7c","Type":"ContainerStarted","Data":"02f695ad4389ce520806101e911edfa960612e3add1aea9538787ea9bf02093b"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.727452 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" event={"ID":"414c531b-a0c8-4cb4-a997-77706c2f8e7c","Type":"ContainerStarted","Data":"9252b862b8115c98b04d770d4c9380993de462adf955805fe0cb2575a21c75df"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.728893 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" podUID="414c531b-a0c8-4cb4-a997-77706c2f8e7c" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.760258 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" event={"ID":"7de7f612-a5bd-4f2c-a146-d5f49ba757af","Type":"ContainerStarted","Data":"8d6f8957c1b6c38d630b0357d8100e304f37031134e869a0c2d704cf7a2f7560"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.760309 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" event={"ID":"7de7f612-a5bd-4f2c-a146-d5f49ba757af","Type":"ContainerStarted","Data":"900f7ecb55bc999b08c55033a51c46adaadc0298f12abdbd5c9fab9d4f0b9990"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.760319 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" event={"ID":"7de7f612-a5bd-4f2c-a146-d5f49ba757af","Type":"ContainerStarted","Data":"c7bd987120c7d8338daea40008965c453b07c1acbafd62f647ac2a9409050f90"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.761005 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.783350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" event={"ID":"426cfad6-045e-48e1-82e4-ce9b9891dd0a","Type":"ContainerStarted","Data":"f255d3b5bd2a0d10657f81899125efd162379a9b0e493c2d74100c616481c99b"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.783400 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" event={"ID":"426cfad6-045e-48e1-82e4-ce9b9891dd0a","Type":"ContainerStarted","Data":"cba0316e2939e808fe649151c6f216f3966bbcde4b3d36422c0105b599840ce4"} Oct 01 07:22:59 crc kubenswrapper[4837]: E1001 07:22:59.790771 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" podUID="426cfad6-045e-48e1-82e4-ce9b9891dd0a" Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.798900 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" event={"ID":"fef6cdf1-8dbe-47e9-8d8a-143710d954ec","Type":"ContainerStarted","Data":"3773fe3e2e01917b935304810353963f2c16d0dbf8325dcd282b3c723f187932"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.823117 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" event={"ID":"5d9a421b-7104-462a-8a01-8bc62522a543","Type":"ContainerStarted","Data":"5928dd1f896b68e5026b82bf3558fb9710effef276dd461bd0ebca98dfd5fa2c"} Oct 01 07:22:59 crc kubenswrapper[4837]: I1001 07:22:59.874988 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" podStartSLOduration=3.8749695109999998 podStartE2EDuration="3.874969511s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:22:59.869105716 +0000 UTC m=+1036.710713171" watchObservedRunningTime="2025-10-01 07:22:59.874969511 +0000 UTC m=+1036.716576966" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.839273 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" podUID="414c531b-a0c8-4cb4-a997-77706c2f8e7c" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.840972 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" podUID="0a3d4dda-cb7c-48ae-95ee-f6eebe98745e" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.841041 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" podUID="969f8b13-b131-4fad-84ff-6b092fcc43e9" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.841087 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" podUID="3efcdf55-46eb-446e-933a-e086f3a67aab" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.841107 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" podUID="421ed6b2-06bf-421e-a871-ec953b32bc93" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.841006 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" podUID="2297d6d0-957e-4d04-9194-d9d7c120c3f0" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.841137 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" podUID="d4fc2a84-086a-4644-b354-2ef743bf1e7f" Oct 01 07:23:00 crc kubenswrapper[4837]: E1001 07:23:00.849846 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" podUID="426cfad6-045e-48e1-82e4-ce9b9891dd0a" Oct 01 07:23:07 crc kubenswrapper[4837]: I1001 07:23:07.925523 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8648d97544-8vjp6" Oct 01 07:23:08 crc kubenswrapper[4837]: I1001 07:23:08.908076 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" event={"ID":"06ebbbd0-256b-4637-b6e1-ba5adab9bd40","Type":"ContainerStarted","Data":"07e98197ebc1fd74b9ab46c11f457a54f8f1a873180c08e8e2fd14d72702b27a"} Oct 01 07:23:08 crc kubenswrapper[4837]: I1001 07:23:08.914047 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" event={"ID":"e5e81392-cf50-43ab-b86a-e8881d5043f7","Type":"ContainerStarted","Data":"571603e857387587f46555c0ea71058b44bb0e1a50b4ca1530643f128399d9e8"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.921130 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" event={"ID":"b537c4f7-697f-4856-b7f9-98a55150cc9f","Type":"ContainerStarted","Data":"fd77a900cb94c42a7f31b33a2d4bc736a80509b3aa30a821a33147fb774f88e9"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.922102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" event={"ID":"b537c4f7-697f-4856-b7f9-98a55150cc9f","Type":"ContainerStarted","Data":"27105f070caab50ae16d6e7c937a1bf4ebc15969c5987ce07483a9938c993cff"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.922186 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.922409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" event={"ID":"f987881b-66b4-49fe-a171-15813e9225f5","Type":"ContainerStarted","Data":"919eb05aef434668768e02f26631b6a27ea527de8dde051d42a04172a6e4a4ab"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.923938 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" event={"ID":"2ed39459-5a70-4c13-aacd-966aae697780","Type":"ContainerStarted","Data":"26e2724f9cd34462dc92c11694ebeac2eb00ccfb602d3d7635544aa8b31272ae"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.925225 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" event={"ID":"e6fb38c5-e91c-4229-9aa5-93729ffab045","Type":"ContainerStarted","Data":"98a36296e9d5df9a10914ac8eaefe4331c100111d25a0ad9e04d417f3d6e1655"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.926714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" event={"ID":"ebdcaccc-170c-40ff-9b06-57a1188cf20b","Type":"ContainerStarted","Data":"52528f3c190d111c345dc13fd7ece497da8e5a621d315948508c7fce65dab3bb"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.928349 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" event={"ID":"750df3b7-ef90-4f83-99f9-ebb54a6a159b","Type":"ContainerStarted","Data":"69679e45e4ba05847683e8e38a4a267d7149cc727231aa63f7de116d907dfd3e"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.928374 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" event={"ID":"750df3b7-ef90-4f83-99f9-ebb54a6a159b","Type":"ContainerStarted","Data":"c3d2155153db51b70412c73b81515ec67a78e1fcd88810a370ad7f16c353f5f6"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.928522 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.930531 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" event={"ID":"648417c5-8d92-4426-81db-052bcb20cbc6","Type":"ContainerStarted","Data":"27ba2d3047cebef6ad26fa30410c8d1975c83c59cc19c7a55d281c1e8eb7fbe1"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.930557 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" event={"ID":"648417c5-8d92-4426-81db-052bcb20cbc6","Type":"ContainerStarted","Data":"66c75a2568775b6b1eaaadf74db930ec81b92dc235445af61d0e9660b86deba6"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.930603 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.932284 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" event={"ID":"5d9a421b-7104-462a-8a01-8bc62522a543","Type":"ContainerStarted","Data":"d32630571db49e61eb3952b53f2a2c220c6c08b81908f3fd5db421890f2485ec"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.933857 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" event={"ID":"e5e81392-cf50-43ab-b86a-e8881d5043f7","Type":"ContainerStarted","Data":"c9700f6b2695f0a3d14dcc0665c5ce050f3e6f53ee9cd486e3ab54f05d19d151"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.933944 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.935449 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" event={"ID":"06ebbbd0-256b-4637-b6e1-ba5adab9bd40","Type":"ContainerStarted","Data":"b66ca56be3c8194ea10f2e09551b3600d37709e26521a056114c43a404e79796"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.935574 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.936941 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" event={"ID":"fef6cdf1-8dbe-47e9-8d8a-143710d954ec","Type":"ContainerStarted","Data":"ea9d3cec72006b1e56836457c9c9b567159ef4bd81c43f461e02cfa5aa70dbfa"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.936979 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" event={"ID":"fef6cdf1-8dbe-47e9-8d8a-143710d954ec","Type":"ContainerStarted","Data":"8907e7a630cd38d43a58f5f65a8eba6cbfe498dd392ac0317800d8a691ad4757"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.937019 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.938215 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" event={"ID":"76e4d2ee-d38f-433c-945c-29f36594a338","Type":"ContainerStarted","Data":"bd19214d21b9885c320bf0920431031ddb51bf42780f6e3edd381bbe33ec0784"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.939553 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" event={"ID":"70c21bc2-f519-4551-bf8d-669df6064834","Type":"ContainerStarted","Data":"0e544489bbdedb05843fbbbc3063fbb2df0f978a7a1e795022f195d9da4d77c3"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.942758 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" event={"ID":"0cd7b250-33bc-45fc-be0d-e4d6d28e450f","Type":"ContainerStarted","Data":"9d1d64cfb886794042e8585f734c353d5178ed2828b582d851bee9db08a1ca4a"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.942790 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" event={"ID":"0cd7b250-33bc-45fc-be0d-e4d6d28e450f","Type":"ContainerStarted","Data":"1f3e6a637d7ab870cbe2dac0acd942375fa92d6e1729e2e068adb9fcd6bd0d41"} Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.942904 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.944042 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" podStartSLOduration=3.992137828 podStartE2EDuration="13.944031939s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.521736432 +0000 UTC m=+1035.363343887" lastFinishedPulling="2025-10-01 07:23:08.473630543 +0000 UTC m=+1045.315237998" observedRunningTime="2025-10-01 07:23:09.940794829 +0000 UTC m=+1046.782402284" watchObservedRunningTime="2025-10-01 07:23:09.944031939 +0000 UTC m=+1046.785639394" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.960434 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" podStartSLOduration=3.388797805 podStartE2EDuration="13.960420493s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.870520979 +0000 UTC m=+1034.712128434" lastFinishedPulling="2025-10-01 07:23:08.442143667 +0000 UTC m=+1045.283751122" observedRunningTime="2025-10-01 07:23:09.956706261 +0000 UTC m=+1046.798313716" watchObservedRunningTime="2025-10-01 07:23:09.960420493 +0000 UTC m=+1046.802027948" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.975249 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" podStartSLOduration=3.448449487 podStartE2EDuration="13.975231617s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.968739412 +0000 UTC m=+1034.810346867" lastFinishedPulling="2025-10-01 07:23:08.495521532 +0000 UTC m=+1045.337128997" observedRunningTime="2025-10-01 07:23:09.971534167 +0000 UTC m=+1046.813141622" watchObservedRunningTime="2025-10-01 07:23:09.975231617 +0000 UTC m=+1046.816839072" Oct 01 07:23:09 crc kubenswrapper[4837]: I1001 07:23:09.987649 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" podStartSLOduration=3.371641962 podStartE2EDuration="13.987635003s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.868678164 +0000 UTC m=+1034.710285619" lastFinishedPulling="2025-10-01 07:23:08.484671205 +0000 UTC m=+1045.326278660" observedRunningTime="2025-10-01 07:23:09.985499471 +0000 UTC m=+1046.827106926" watchObservedRunningTime="2025-10-01 07:23:09.987635003 +0000 UTC m=+1046.829242458" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.013494 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" podStartSLOduration=4.143651701 podStartE2EDuration="14.01347645s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.614804445 +0000 UTC m=+1035.456411900" lastFinishedPulling="2025-10-01 07:23:08.484629184 +0000 UTC m=+1045.326236649" observedRunningTime="2025-10-01 07:23:10.011790068 +0000 UTC m=+1046.853397523" watchObservedRunningTime="2025-10-01 07:23:10.01347645 +0000 UTC m=+1046.855083905" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.035901 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" podStartSLOduration=4.195642402 podStartE2EDuration="14.035886612s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.65357788 +0000 UTC m=+1035.495185335" lastFinishedPulling="2025-10-01 07:23:08.49382209 +0000 UTC m=+1045.335429545" observedRunningTime="2025-10-01 07:23:10.031881153 +0000 UTC m=+1046.873488608" watchObservedRunningTime="2025-10-01 07:23:10.035886612 +0000 UTC m=+1046.877494067" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.058669 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" podStartSLOduration=3.531470353 podStartE2EDuration="14.058654513s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.972947086 +0000 UTC m=+1034.814554541" lastFinishedPulling="2025-10-01 07:23:08.500131246 +0000 UTC m=+1045.341738701" observedRunningTime="2025-10-01 07:23:10.05652168 +0000 UTC m=+1046.898129135" watchObservedRunningTime="2025-10-01 07:23:10.058654513 +0000 UTC m=+1046.900261968" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.954271 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" event={"ID":"76e4d2ee-d38f-433c-945c-29f36594a338","Type":"ContainerStarted","Data":"e322fa4fca8ee3be0b945d73fb572fbd41206515d5e01571aa2c21a29cb61d5e"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.954662 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.957290 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" event={"ID":"70c21bc2-f519-4551-bf8d-669df6064834","Type":"ContainerStarted","Data":"1f252fba20de0499902bcfa8cb9c139e32db39c36bd091728e218dd786939713"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.957457 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.959344 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" event={"ID":"2ed39459-5a70-4c13-aacd-966aae697780","Type":"ContainerStarted","Data":"bc369cb6e95d32946cea76196b19f717c69b0c6579e591cfd1f6fd6aefabc77b"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.959565 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.963448 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" event={"ID":"ebdcaccc-170c-40ff-9b06-57a1188cf20b","Type":"ContainerStarted","Data":"9c86060251590f44b565ff6588e1c7f8c015901cdfadb02cf0461b0a6722cfff"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.963626 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.966280 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" event={"ID":"5d9a421b-7104-462a-8a01-8bc62522a543","Type":"ContainerStarted","Data":"5a9a8b550f41776bfd3b51e32c59335a45e6b5a4ee9081627cd453d2167942f4"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.970767 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" event={"ID":"f987881b-66b4-49fe-a171-15813e9225f5","Type":"ContainerStarted","Data":"e7287a53ea637c3e7ff3f66fc65f2b5fe7bafea725f5c5bdfcb38bb9745b1c0a"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.971662 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.985090 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" event={"ID":"e6fb38c5-e91c-4229-9aa5-93729ffab045","Type":"ContainerStarted","Data":"dfa33d5ea261836cdc60dc5373012629e484167efde23943511ac0052198f9d1"} Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.985150 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:23:10 crc kubenswrapper[4837]: I1001 07:23:10.986869 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" podStartSLOduration=5.015080739 podStartE2EDuration="14.98684791s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.522131621 +0000 UTC m=+1035.363739076" lastFinishedPulling="2025-10-01 07:23:08.493898792 +0000 UTC m=+1045.335506247" observedRunningTime="2025-10-01 07:23:10.9832096 +0000 UTC m=+1047.824817085" watchObservedRunningTime="2025-10-01 07:23:10.98684791 +0000 UTC m=+1047.828455405" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.021124 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" podStartSLOduration=5.16605821 podStartE2EDuration="15.02092784s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.613661087 +0000 UTC m=+1035.455268542" lastFinishedPulling="2025-10-01 07:23:08.468530717 +0000 UTC m=+1045.310138172" observedRunningTime="2025-10-01 07:23:11.013498977 +0000 UTC m=+1047.855106472" watchObservedRunningTime="2025-10-01 07:23:11.02092784 +0000 UTC m=+1047.862535335" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.036945 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" podStartSLOduration=4.520207632 podStartE2EDuration="15.036922824s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.976154115 +0000 UTC m=+1034.817761570" lastFinishedPulling="2025-10-01 07:23:08.492869307 +0000 UTC m=+1045.334476762" observedRunningTime="2025-10-01 07:23:11.028948077 +0000 UTC m=+1047.870555552" watchObservedRunningTime="2025-10-01 07:23:11.036922824 +0000 UTC m=+1047.878530319" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.051640 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" podStartSLOduration=4.380596274 podStartE2EDuration="15.051621146s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:57.825116235 +0000 UTC m=+1034.666723680" lastFinishedPulling="2025-10-01 07:23:08.496141087 +0000 UTC m=+1045.337748552" observedRunningTime="2025-10-01 07:23:11.047624587 +0000 UTC m=+1047.889232072" watchObservedRunningTime="2025-10-01 07:23:11.051621146 +0000 UTC m=+1047.893228611" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.077225 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" podStartSLOduration=5.117161666 podStartE2EDuration="15.077195046s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.569082149 +0000 UTC m=+1035.410689604" lastFinishedPulling="2025-10-01 07:23:08.529115509 +0000 UTC m=+1045.370722984" observedRunningTime="2025-10-01 07:23:11.073810053 +0000 UTC m=+1047.915417548" watchObservedRunningTime="2025-10-01 07:23:11.077195046 +0000 UTC m=+1047.918802531" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.093139 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" podStartSLOduration=5.195982247 podStartE2EDuration="15.093113949s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.602711847 +0000 UTC m=+1035.444319302" lastFinishedPulling="2025-10-01 07:23:08.499843539 +0000 UTC m=+1045.341451004" observedRunningTime="2025-10-01 07:23:11.092115554 +0000 UTC m=+1047.933723049" watchObservedRunningTime="2025-10-01 07:23:11.093113949 +0000 UTC m=+1047.934721444" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.115537 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" podStartSLOduration=5.222585981 podStartE2EDuration="15.11551627s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.60690512 +0000 UTC m=+1035.448512575" lastFinishedPulling="2025-10-01 07:23:08.499835399 +0000 UTC m=+1045.341442864" observedRunningTime="2025-10-01 07:23:11.107918883 +0000 UTC m=+1047.949526378" watchObservedRunningTime="2025-10-01 07:23:11.11551627 +0000 UTC m=+1047.957123765" Oct 01 07:23:11 crc kubenswrapper[4837]: I1001 07:23:11.992858 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.566583 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-rqr6f" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.581497 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-lm9xd" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.610060 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-wnqvm" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.613142 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-lq7sr" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.668746 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-rz6jg" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.697481 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f589bc7f7-wf96x" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.756634 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-mgm24" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.789282 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6b96467f46-kgf8r" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.802552 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79f9fc9fd8-p978w" Oct 01 07:23:16 crc kubenswrapper[4837]: I1001 07:23:16.814068 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fb7d6b8bf-2xtdw" Oct 01 07:23:17 crc kubenswrapper[4837]: I1001 07:23:17.005824 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-657c6b68c7-fddfz" Oct 01 07:23:17 crc kubenswrapper[4837]: I1001 07:23:17.029414 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-6hgks" Oct 01 07:23:17 crc kubenswrapper[4837]: I1001 07:23:17.294422 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-tlhpd" Oct 01 07:23:17 crc kubenswrapper[4837]: I1001 07:23:17.600036 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.070167 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" event={"ID":"969f8b13-b131-4fad-84ff-6b092fcc43e9","Type":"ContainerStarted","Data":"4c41201160416239d433be36a31adff62e9edb79500c051a88482f196b0e40a7"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.071059 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.072548 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" event={"ID":"2297d6d0-957e-4d04-9194-d9d7c120c3f0","Type":"ContainerStarted","Data":"a1ec684267c27724facd5edcbf339d4a58ff302bfc055bda529a479ba9850ad2"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.072828 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.075348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" event={"ID":"d4fc2a84-086a-4644-b354-2ef743bf1e7f","Type":"ContainerStarted","Data":"390c597fe6a429e4222bf039a061398df55ca4ccbd8be83713372e38417031f0"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.075551 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.077837 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" event={"ID":"414c531b-a0c8-4cb4-a997-77706c2f8e7c","Type":"ContainerStarted","Data":"3c50dd586ae9ce9b5fae7713c2b1ca13b7d26cdcd5aa5a26fb2b123f576e2b6e"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.078155 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.081074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" event={"ID":"426cfad6-045e-48e1-82e4-ce9b9891dd0a","Type":"ContainerStarted","Data":"03dd940546dc5452b13e5a133b7761f5310023293a457558918b70670ad8909d"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.081391 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.083286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" event={"ID":"421ed6b2-06bf-421e-a871-ec953b32bc93","Type":"ContainerStarted","Data":"d255b5ff2cdf74596878c0bed67356f5d8b3090c9928862767978898e168c746"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.085587 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" event={"ID":"0a3d4dda-cb7c-48ae-95ee-f6eebe98745e","Type":"ContainerStarted","Data":"b3436fde7dad55428e1b8477cd91f5a3a02362b91ffd17fff1f7443b30fcd3b3"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.086170 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.089433 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" event={"ID":"3efcdf55-46eb-446e-933a-e086f3a67aab","Type":"ContainerStarted","Data":"794368d4a6067a514e4e36987526920a9ca432958146151c0087dcabe345baae"} Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.089815 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.102935 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" podStartSLOduration=4.046500439 podStartE2EDuration="24.102909361s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.732916135 +0000 UTC m=+1035.574523590" lastFinishedPulling="2025-10-01 07:23:18.789325057 +0000 UTC m=+1055.630932512" observedRunningTime="2025-10-01 07:23:20.086717281 +0000 UTC m=+1056.928324766" watchObservedRunningTime="2025-10-01 07:23:20.102909361 +0000 UTC m=+1056.944516846" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.112978 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" podStartSLOduration=4.061598429 podStartE2EDuration="24.112957768s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.736357259 +0000 UTC m=+1035.577964714" lastFinishedPulling="2025-10-01 07:23:18.787716598 +0000 UTC m=+1055.629324053" observedRunningTime="2025-10-01 07:23:20.106223572 +0000 UTC m=+1056.947831037" watchObservedRunningTime="2025-10-01 07:23:20.112957768 +0000 UTC m=+1056.954565233" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.126751 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" podStartSLOduration=4.037366133 podStartE2EDuration="24.126732478s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.736189675 +0000 UTC m=+1035.577797130" lastFinishedPulling="2025-10-01 07:23:18.82555598 +0000 UTC m=+1055.667163475" observedRunningTime="2025-10-01 07:23:20.121258833 +0000 UTC m=+1056.962866298" watchObservedRunningTime="2025-10-01 07:23:20.126732478 +0000 UTC m=+1056.968339953" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.141956 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2" podStartSLOduration=4.08841726 podStartE2EDuration="24.141932111s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.732914015 +0000 UTC m=+1035.574521470" lastFinishedPulling="2025-10-01 07:23:18.786428866 +0000 UTC m=+1055.628036321" observedRunningTime="2025-10-01 07:23:20.137433701 +0000 UTC m=+1056.979041156" watchObservedRunningTime="2025-10-01 07:23:20.141932111 +0000 UTC m=+1056.983539596" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.162187 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" podStartSLOduration=3.993123692 podStartE2EDuration="24.16216676s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.619216864 +0000 UTC m=+1035.460824319" lastFinishedPulling="2025-10-01 07:23:18.788259922 +0000 UTC m=+1055.629867387" observedRunningTime="2025-10-01 07:23:20.158448998 +0000 UTC m=+1057.000056483" watchObservedRunningTime="2025-10-01 07:23:20.16216676 +0000 UTC m=+1057.003774215" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.178322 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" podStartSLOduration=4.064206033 podStartE2EDuration="24.178301267s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.666267493 +0000 UTC m=+1035.507874948" lastFinishedPulling="2025-10-01 07:23:18.780362737 +0000 UTC m=+1055.621970182" observedRunningTime="2025-10-01 07:23:20.172517455 +0000 UTC m=+1057.014124910" watchObservedRunningTime="2025-10-01 07:23:20.178301267 +0000 UTC m=+1057.019908732" Oct 01 07:23:20 crc kubenswrapper[4837]: I1001 07:23:20.196174 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" podStartSLOduration=4.062968862 podStartE2EDuration="24.196153157s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.653660812 +0000 UTC m=+1035.495268267" lastFinishedPulling="2025-10-01 07:23:18.786845107 +0000 UTC m=+1055.628452562" observedRunningTime="2025-10-01 07:23:20.188870968 +0000 UTC m=+1057.030478443" watchObservedRunningTime="2025-10-01 07:23:20.196153157 +0000 UTC m=+1057.037760622" Oct 01 07:23:23 crc kubenswrapper[4837]: I1001 07:23:23.080223 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:23:23 crc kubenswrapper[4837]: I1001 07:23:23.080642 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:23:26 crc kubenswrapper[4837]: I1001 07:23:26.897202 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-6bm6v" Oct 01 07:23:26 crc kubenswrapper[4837]: I1001 07:23:26.929845 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" podStartSLOduration=10.83896917 podStartE2EDuration="30.929820732s" podCreationTimestamp="2025-10-01 07:22:56 +0000 UTC" firstStartedPulling="2025-10-01 07:22:58.736242406 +0000 UTC m=+1035.577849861" lastFinishedPulling="2025-10-01 07:23:18.827093968 +0000 UTC m=+1055.668701423" observedRunningTime="2025-10-01 07:23:20.219667867 +0000 UTC m=+1057.061275362" watchObservedRunningTime="2025-10-01 07:23:26.929820732 +0000 UTC m=+1063.771428227" Oct 01 07:23:26 crc kubenswrapper[4837]: I1001 07:23:26.936877 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-95dns" Oct 01 07:23:26 crc kubenswrapper[4837]: I1001 07:23:26.936949 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-822wp" Oct 01 07:23:27 crc kubenswrapper[4837]: I1001 07:23:27.025869 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-cmqwh" Oct 01 07:23:27 crc kubenswrapper[4837]: I1001 07:23:27.041917 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-l4pdl" Oct 01 07:23:27 crc kubenswrapper[4837]: I1001 07:23:27.077811 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb97fcf96-vcm95" Oct 01 07:23:27 crc kubenswrapper[4837]: I1001 07:23:27.143451 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75756dd4d9-vx2m2" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.420615 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.422569 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.425547 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.425738 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.426027 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.426124 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-m7cmq" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.438304 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.480216 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.481961 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.485949 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.488632 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.565366 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.565412 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.565442 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p52m2\" (UniqueName: \"kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.565482 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.565509 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jmnk\" (UniqueName: \"kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.668568 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.670868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jmnk\" (UniqueName: \"kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.671076 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.671174 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.671274 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p52m2\" (UniqueName: \"kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.672271 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.675111 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.675744 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.688039 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jmnk\" (UniqueName: \"kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk\") pod \"dnsmasq-dns-d5f6f49c7-npcvx\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.694150 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p52m2\" (UniqueName: \"kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2\") pod \"dnsmasq-dns-b8b69cf79-wwhr9\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.739237 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:23:42 crc kubenswrapper[4837]: I1001 07:23:42.795758 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:23:43 crc kubenswrapper[4837]: W1001 07:23:43.242502 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea9623ce_644c_4926_a266_a58914af2c7b.slice/crio-a77a4453cf557135012ba8bc46d68908b6ae75b7ea8a30d812761eb541828e91 WatchSource:0}: Error finding container a77a4453cf557135012ba8bc46d68908b6ae75b7ea8a30d812761eb541828e91: Status 404 returned error can't find the container with id a77a4453cf557135012ba8bc46d68908b6ae75b7ea8a30d812761eb541828e91 Oct 01 07:23:43 crc kubenswrapper[4837]: I1001 07:23:43.244123 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:23:43 crc kubenswrapper[4837]: I1001 07:23:43.246859 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:23:43 crc kubenswrapper[4837]: I1001 07:23:43.278416 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:23:43 crc kubenswrapper[4837]: W1001 07:23:43.280414 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a11892_499b_4008_8c40_01a979abe951.slice/crio-e89cf59e1ffea718f1c6ca73bf2a4d642106976f02c5b7b4c7f98f047224cbbb WatchSource:0}: Error finding container e89cf59e1ffea718f1c6ca73bf2a4d642106976f02c5b7b4c7f98f047224cbbb: Status 404 returned error can't find the container with id e89cf59e1ffea718f1c6ca73bf2a4d642106976f02c5b7b4c7f98f047224cbbb Oct 01 07:23:43 crc kubenswrapper[4837]: I1001 07:23:43.318487 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" event={"ID":"51a11892-499b-4008-8c40-01a979abe951","Type":"ContainerStarted","Data":"e89cf59e1ffea718f1c6ca73bf2a4d642106976f02c5b7b4c7f98f047224cbbb"} Oct 01 07:23:43 crc kubenswrapper[4837]: I1001 07:23:43.320115 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" event={"ID":"ea9623ce-644c-4926-a266-a58914af2c7b","Type":"ContainerStarted","Data":"a77a4453cf557135012ba8bc46d68908b6ae75b7ea8a30d812761eb541828e91"} Oct 01 07:23:44 crc kubenswrapper[4837]: I1001 07:23:44.864759 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:23:44 crc kubenswrapper[4837]: I1001 07:23:44.882927 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:23:44 crc kubenswrapper[4837]: I1001 07:23:44.886086 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:44 crc kubenswrapper[4837]: I1001 07:23:44.890380 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.004652 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.004718 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.004744 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9kcs\" (UniqueName: \"kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.109507 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.109560 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.109583 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9kcs\" (UniqueName: \"kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.110576 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.111106 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.154766 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9kcs\" (UniqueName: \"kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs\") pod \"dnsmasq-dns-b6f94bdfc-nfc9g\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.184843 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.227530 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.228296 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.229058 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.233092 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.318792 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bj6\" (UniqueName: \"kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.319128 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.319202 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.421400 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.421465 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bj6\" (UniqueName: \"kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.421494 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.422311 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.422335 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.454320 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bj6\" (UniqueName: \"kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6\") pod \"dnsmasq-dns-77795d58f5-hcbzv\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.557478 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.762932 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:23:45 crc kubenswrapper[4837]: W1001 07:23:45.788173 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaf9c3c5_dc37_415d_bed9_33af89fffd07.slice/crio-13fe0f27614d0c26837f83bb6e0f727336fd2761e5c4f72f9ac2682d945328c2 WatchSource:0}: Error finding container 13fe0f27614d0c26837f83bb6e0f727336fd2761e5c4f72f9ac2682d945328c2: Status 404 returned error can't find the container with id 13fe0f27614d0c26837f83bb6e0f727336fd2761e5c4f72f9ac2682d945328c2 Oct 01 07:23:45 crc kubenswrapper[4837]: I1001 07:23:45.996481 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:23:46 crc kubenswrapper[4837]: W1001 07:23:46.003164 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a215cd7_0557_42f0_96b0_8ea079026f6b.slice/crio-a479f71972b0ac246c9858d14e26af52009b872521a5ef05568f002bdfe68306 WatchSource:0}: Error finding container a479f71972b0ac246c9858d14e26af52009b872521a5ef05568f002bdfe68306: Status 404 returned error can't find the container with id a479f71972b0ac246c9858d14e26af52009b872521a5ef05568f002bdfe68306 Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.030042 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.031278 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.034731 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.034771 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.034917 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.035135 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.035285 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.035481 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.035683 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2b9gt" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.041566 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130538 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130567 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130586 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130650 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130729 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.130759 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.131098 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.131431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.131480 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.131540 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5vz4\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232499 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232544 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232594 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232627 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232676 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232710 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232725 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5vz4\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232752 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232766 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232781 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.232795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.234024 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.234501 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.234797 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.236214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.236291 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.236543 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.244352 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.244400 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.250297 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.250337 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.254193 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5vz4\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.259367 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.352744 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.358783 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.362020 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.363507 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" event={"ID":"aaf9c3c5-dc37-415d-bed9-33af89fffd07","Type":"ContainerStarted","Data":"13fe0f27614d0c26837f83bb6e0f727336fd2761e5c4f72f9ac2682d945328c2"} Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366029 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366214 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366223 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366360 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366508 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366678 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.366797 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xmg68" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.368308 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.371064 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" event={"ID":"3a215cd7-0557-42f0-96b0-8ea079026f6b","Type":"ContainerStarted","Data":"a479f71972b0ac246c9858d14e26af52009b872521a5ef05568f002bdfe68306"} Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436439 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436777 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436817 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436843 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436900 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436947 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxdj2\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436966 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.436986 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.437009 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.437204 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.538909 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.538965 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.538990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539024 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539061 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539093 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxdj2\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539133 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539152 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.539180 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.540357 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.540742 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.540967 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.541108 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.541183 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.542222 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.543034 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.545015 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.545528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.558134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.560083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxdj2\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.561320 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:46 crc kubenswrapper[4837]: I1001 07:23:46.698406 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.953292 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.954933 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.959041 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.959053 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.959205 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-j7gvb" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.959292 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.960015 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 01 07:23:47 crc kubenswrapper[4837]: I1001 07:23:47.972283 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.002089 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075374 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075419 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075458 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075482 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075518 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075538 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075551 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g85kd\" (UniqueName: \"kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075573 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.075616 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.176900 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.176947 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.176990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177025 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g85kd\" (UniqueName: \"kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177050 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177097 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177114 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177129 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177234 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177297 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.177960 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.178018 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.178507 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.181795 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.182006 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.202571 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.203210 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g85kd\" (UniqueName: \"kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.225349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.271101 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.964622 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.967858 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.970478 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jqfkj" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.970863 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.977525 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.977875 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 01 07:23:48 crc kubenswrapper[4837]: I1001 07:23:48.985451 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092111 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092225 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092363 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092421 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092502 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092537 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092563 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97dld\" (UniqueName: \"kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092592 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.092634 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194151 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194207 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97dld\" (UniqueName: \"kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194234 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194268 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194325 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194369 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194400 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194431 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194478 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.194868 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.195157 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.195415 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.195486 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.196246 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.198134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.210194 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.213817 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.214955 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97dld\" (UniqueName: \"kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.233132 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.275393 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.276582 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.282148 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.282189 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.282411 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-sbxqc" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.288759 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.295541 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.295589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.295939 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.296040 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz82z\" (UniqueName: \"kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.296162 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.309743 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.396859 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.396905 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz82z\" (UniqueName: \"kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.396942 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.396973 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.396990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.397745 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.397801 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.418373 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.418414 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.424319 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz82z\" (UniqueName: \"kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z\") pod \"memcached-0\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " pod="openstack/memcached-0" Oct 01 07:23:49 crc kubenswrapper[4837]: I1001 07:23:49.602851 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 07:23:50 crc kubenswrapper[4837]: I1001 07:23:50.276904 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.126784 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.127738 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.136405 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.142502 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-km925" Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.223287 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q76hj\" (UniqueName: \"kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj\") pod \"kube-state-metrics-0\" (UID: \"14422052-034a-4e59-a29b-6be7aa9d077b\") " pod="openstack/kube-state-metrics-0" Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.324941 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q76hj\" (UniqueName: \"kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj\") pod \"kube-state-metrics-0\" (UID: \"14422052-034a-4e59-a29b-6be7aa9d077b\") " pod="openstack/kube-state-metrics-0" Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.348737 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q76hj\" (UniqueName: \"kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj\") pod \"kube-state-metrics-0\" (UID: \"14422052-034a-4e59-a29b-6be7aa9d077b\") " pod="openstack/kube-state-metrics-0" Oct 01 07:23:51 crc kubenswrapper[4837]: I1001 07:23:51.444143 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:23:53 crc kubenswrapper[4837]: I1001 07:23:53.079972 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:23:53 crc kubenswrapper[4837]: I1001 07:23:53.080435 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:23:53 crc kubenswrapper[4837]: W1001 07:23:53.111902 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dcde73a_d928_45af_914d_3b1f5f7f899b.slice/crio-6e4114f7e83256f7322cb3d583e2616219837e99d678ca299638e4b02d3305aa WatchSource:0}: Error finding container 6e4114f7e83256f7322cb3d583e2616219837e99d678ca299638e4b02d3305aa: Status 404 returned error can't find the container with id 6e4114f7e83256f7322cb3d583e2616219837e99d678ca299638e4b02d3305aa Oct 01 07:23:53 crc kubenswrapper[4837]: I1001 07:23:53.470998 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerStarted","Data":"6e4114f7e83256f7322cb3d583e2616219837e99d678ca299638e4b02d3305aa"} Oct 01 07:23:53 crc kubenswrapper[4837]: I1001 07:23:53.557992 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.145616 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.146840 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.149046 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mhnlb" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.149061 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.149775 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.150019 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.153724 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.168054 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.225178 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.226295 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.229893 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.230091 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.230203 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-9kjr9" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.245335 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275679 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275733 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275762 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275860 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275878 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275903 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqtz4\" (UniqueName: \"kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.275935 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.290629 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.293856 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.305079 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377340 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377393 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377443 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tbcv\" (UniqueName: \"kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377476 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377496 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377518 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377592 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377635 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377705 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377734 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377758 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.377789 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqtz4\" (UniqueName: \"kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.378370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.378663 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.378964 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.379419 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.384436 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.393899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.394978 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqtz4\" (UniqueName: \"kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.405034 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.410957 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.463476 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478547 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478600 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478643 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478671 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tbcv\" (UniqueName: \"kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478705 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478727 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478745 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkcp4\" (UniqueName: \"kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478776 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478823 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478852 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478878 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.478896 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.479396 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.479513 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.479573 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.480846 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.482299 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.482971 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.495316 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tbcv\" (UniqueName: \"kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv\") pod \"ovn-controller-kt6jr\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.546353 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580430 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580531 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580575 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580611 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkcp4\" (UniqueName: \"kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.580643 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.584022 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.584165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.584222 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.584224 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.584642 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.604126 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkcp4\" (UniqueName: \"kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4\") pod \"ovn-controller-ovs-zpx7h\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:55 crc kubenswrapper[4837]: I1001 07:23:55.630208 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.661845 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.667127 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.670718 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.670972 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.671118 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.674279 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-gzg2g" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.675711 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.830883 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.830967 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831047 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48msk\" (UniqueName: \"kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831085 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831131 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831162 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831201 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.831287 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.932733 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.932883 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.932933 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.933003 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48msk\" (UniqueName: \"kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.933090 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.933790 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.933820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.933871 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.934982 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.935015 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.935423 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.935948 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.938970 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.945430 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.947121 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.957259 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48msk\" (UniqueName: \"kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:58 crc kubenswrapper[4837]: I1001 07:23:58.968407 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:59 crc kubenswrapper[4837]: I1001 07:23:59.003861 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 07:23:59 crc kubenswrapper[4837]: I1001 07:23:59.389405 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:23:59 crc kubenswrapper[4837]: I1001 07:23:59.511225 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerStarted","Data":"eb4b6ed36fb446c04f97e906a0e9933f49802e8c4266043a294f9ccd3aeeb126"} Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.017232 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b" Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.017616 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p52m2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-b8b69cf79-wwhr9_openstack(ea9623ce-644c-4926-a266-a58914af2c7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.019472 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" podUID="ea9623ce-644c-4926-a266-a58914af2c7b" Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.047851 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b" Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.048032 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5jmnk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-d5f6f49c7-npcvx_openstack(51a11892-499b-4008-8c40-01a979abe951): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 07:24:00 crc kubenswrapper[4837]: E1001 07:24:00.049351 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" podUID="51a11892-499b-4008-8c40-01a979abe951" Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.446397 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.521661 4837 generic.go:334] "Generic (PLEG): container finished" podID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerID="7607a88a63c494b3dc13e2e72ff3705ae72892ea40bd984c45d3543fb5957bf0" exitCode=0 Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.521724 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" event={"ID":"aaf9c3c5-dc37-415d-bed9-33af89fffd07","Type":"ContainerDied","Data":"7607a88a63c494b3dc13e2e72ff3705ae72892ea40bd984c45d3543fb5957bf0"} Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.524974 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7f2490fe-759d-4263-ae63-0f69c4f78c7d","Type":"ContainerStarted","Data":"3c99a37f87139fe330d27863ab221dcb31ac8a65601421c6d91492703b1831c6"} Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.534538 4837 generic.go:334] "Generic (PLEG): container finished" podID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerID="8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39" exitCode=0 Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.534625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" event={"ID":"3a215cd7-0557-42f0-96b0-8ea079026f6b","Type":"ContainerDied","Data":"8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39"} Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.545867 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.548132 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerStarted","Data":"dc29fcfd5af62b329b544cd0c06967919cdfbcde1b023e88bda0c6c9ccecb870"} Oct 01 07:24:00 crc kubenswrapper[4837]: W1001 07:24:00.575842 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44374bf9_c936_4c06_ac74_e80244c09039.slice/crio-fc4268e9129deeb80bfeefd6bd5daedba90027177c86e100dc825a945316fd84 WatchSource:0}: Error finding container fc4268e9129deeb80bfeefd6bd5daedba90027177c86e100dc825a945316fd84: Status 404 returned error can't find the container with id fc4268e9129deeb80bfeefd6bd5daedba90027177c86e100dc825a945316fd84 Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.709621 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.729833 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.845330 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.905395 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:24:00 crc kubenswrapper[4837]: I1001 07:24:00.972101 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.086362 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.092031 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.173834 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jmnk\" (UniqueName: \"kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk\") pod \"51a11892-499b-4008-8c40-01a979abe951\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.173884 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config\") pod \"ea9623ce-644c-4926-a266-a58914af2c7b\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.173995 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc\") pod \"51a11892-499b-4008-8c40-01a979abe951\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.174023 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config\") pod \"51a11892-499b-4008-8c40-01a979abe951\" (UID: \"51a11892-499b-4008-8c40-01a979abe951\") " Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.174042 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p52m2\" (UniqueName: \"kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2\") pod \"ea9623ce-644c-4926-a266-a58914af2c7b\" (UID: \"ea9623ce-644c-4926-a266-a58914af2c7b\") " Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.176281 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config" (OuterVolumeSpecName: "config") pod "51a11892-499b-4008-8c40-01a979abe951" (UID: "51a11892-499b-4008-8c40-01a979abe951"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.176290 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51a11892-499b-4008-8c40-01a979abe951" (UID: "51a11892-499b-4008-8c40-01a979abe951"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.176290 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config" (OuterVolumeSpecName: "config") pod "ea9623ce-644c-4926-a266-a58914af2c7b" (UID: "ea9623ce-644c-4926-a266-a58914af2c7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.179942 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk" (OuterVolumeSpecName: "kube-api-access-5jmnk") pod "51a11892-499b-4008-8c40-01a979abe951" (UID: "51a11892-499b-4008-8c40-01a979abe951"). InnerVolumeSpecName "kube-api-access-5jmnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.182520 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2" (OuterVolumeSpecName: "kube-api-access-p52m2") pod "ea9623ce-644c-4926-a266-a58914af2c7b" (UID: "ea9623ce-644c-4926-a266-a58914af2c7b"). InnerVolumeSpecName "kube-api-access-p52m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.278089 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.278119 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a11892-499b-4008-8c40-01a979abe951-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.278129 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p52m2\" (UniqueName: \"kubernetes.io/projected/ea9623ce-644c-4926-a266-a58914af2c7b-kube-api-access-p52m2\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.278138 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jmnk\" (UniqueName: \"kubernetes.io/projected/51a11892-499b-4008-8c40-01a979abe951-kube-api-access-5jmnk\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.278149 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea9623ce-644c-4926-a266-a58914af2c7b-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.559159 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerStarted","Data":"0a85e84c711ab07e9e92ee43b138119648100285a03a48d1d7e39d9624c83e1a"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.560656 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr" event={"ID":"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62","Type":"ContainerStarted","Data":"c847295e489a5acc261e47c3c1388e5e6983c1f8d86bc2a53ea898868e3d233d"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.562803 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerStarted","Data":"cae936c3d8781371574591fd0f37891f1b579b4a93242b367bfd809455712aec"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.564168 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerStarted","Data":"ed403b18b72ad5bfcd988a39cf2a3cbcc8a9c28a2048534a8fa96bb821df130b"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.566941 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.566946 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5f6f49c7-npcvx" event={"ID":"51a11892-499b-4008-8c40-01a979abe951","Type":"ContainerDied","Data":"e89cf59e1ffea718f1c6ca73bf2a4d642106976f02c5b7b4c7f98f047224cbbb"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.570228 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" event={"ID":"3a215cd7-0557-42f0-96b0-8ea079026f6b","Type":"ContainerStarted","Data":"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.570344 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.572793 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" event={"ID":"aaf9c3c5-dc37-415d-bed9-33af89fffd07","Type":"ContainerStarted","Data":"05a9fe216ab76787903f88d032dc165a8ec22819a3f02939ea40e285fdbb5850"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.573296 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.574484 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerStarted","Data":"fc4268e9129deeb80bfeefd6bd5daedba90027177c86e100dc825a945316fd84"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.576537 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"14422052-034a-4e59-a29b-6be7aa9d077b","Type":"ContainerStarted","Data":"fd4306543ebd110b5209a1acde5f8b9d81a4e18a43c3efb841eb4f85d0ac3f70"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.588022 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" event={"ID":"ea9623ce-644c-4926-a266-a58914af2c7b","Type":"ContainerDied","Data":"a77a4453cf557135012ba8bc46d68908b6ae75b7ea8a30d812761eb541828e91"} Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.588080 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-wwhr9" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.588378 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" podStartSLOduration=2.407197978 podStartE2EDuration="16.588368007s" podCreationTimestamp="2025-10-01 07:23:45 +0000 UTC" firstStartedPulling="2025-10-01 07:23:46.005129466 +0000 UTC m=+1082.846736921" lastFinishedPulling="2025-10-01 07:24:00.186299495 +0000 UTC m=+1097.027906950" observedRunningTime="2025-10-01 07:24:01.587936296 +0000 UTC m=+1098.429543761" watchObservedRunningTime="2025-10-01 07:24:01.588368007 +0000 UTC m=+1098.429975462" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.608043 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" podStartSLOduration=3.148949475 podStartE2EDuration="17.608026861s" podCreationTimestamp="2025-10-01 07:23:44 +0000 UTC" firstStartedPulling="2025-10-01 07:23:45.79100573 +0000 UTC m=+1082.632613185" lastFinishedPulling="2025-10-01 07:24:00.250083116 +0000 UTC m=+1097.091690571" observedRunningTime="2025-10-01 07:24:01.602536606 +0000 UTC m=+1098.444144061" watchObservedRunningTime="2025-10-01 07:24:01.608026861 +0000 UTC m=+1098.449634316" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.652927 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.656707 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-npcvx"] Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.677567 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.682132 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-wwhr9"] Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.825018 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a11892-499b-4008-8c40-01a979abe951" path="/var/lib/kubelet/pods/51a11892-499b-4008-8c40-01a979abe951/volumes" Oct 01 07:24:01 crc kubenswrapper[4837]: I1001 07:24:01.825351 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9623ce-644c-4926-a266-a58914af2c7b" path="/var/lib/kubelet/pods/ea9623ce-644c-4926-a266-a58914af2c7b/volumes" Oct 01 07:24:05 crc kubenswrapper[4837]: I1001 07:24:05.230887 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:24:05 crc kubenswrapper[4837]: I1001 07:24:05.558846 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:24:05 crc kubenswrapper[4837]: I1001 07:24:05.603331 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:24:05 crc kubenswrapper[4837]: I1001 07:24:05.624293 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="dnsmasq-dns" containerID="cri-o://05a9fe216ab76787903f88d032dc165a8ec22819a3f02939ea40e285fdbb5850" gracePeriod=10 Oct 01 07:24:06 crc kubenswrapper[4837]: I1001 07:24:06.633222 4837 generic.go:334] "Generic (PLEG): container finished" podID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerID="05a9fe216ab76787903f88d032dc165a8ec22819a3f02939ea40e285fdbb5850" exitCode=0 Oct 01 07:24:06 crc kubenswrapper[4837]: I1001 07:24:06.633263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" event={"ID":"aaf9c3c5-dc37-415d-bed9-33af89fffd07","Type":"ContainerDied","Data":"05a9fe216ab76787903f88d032dc165a8ec22819a3f02939ea40e285fdbb5850"} Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.842654 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.898900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config\") pod \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.899008 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc\") pod \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.899060 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9kcs\" (UniqueName: \"kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs\") pod \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\" (UID: \"aaf9c3c5-dc37-415d-bed9-33af89fffd07\") " Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.903453 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs" (OuterVolumeSpecName: "kube-api-access-b9kcs") pod "aaf9c3c5-dc37-415d-bed9-33af89fffd07" (UID: "aaf9c3c5-dc37-415d-bed9-33af89fffd07"). InnerVolumeSpecName "kube-api-access-b9kcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.937673 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aaf9c3c5-dc37-415d-bed9-33af89fffd07" (UID: "aaf9c3c5-dc37-415d-bed9-33af89fffd07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:07 crc kubenswrapper[4837]: I1001 07:24:07.949800 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config" (OuterVolumeSpecName: "config") pod "aaf9c3c5-dc37-415d-bed9-33af89fffd07" (UID: "aaf9c3c5-dc37-415d-bed9-33af89fffd07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.000318 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9kcs\" (UniqueName: \"kubernetes.io/projected/aaf9c3c5-dc37-415d-bed9-33af89fffd07-kube-api-access-b9kcs\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.000353 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.000364 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaf9c3c5-dc37-415d-bed9-33af89fffd07-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.653537 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" event={"ID":"aaf9c3c5-dc37-415d-bed9-33af89fffd07","Type":"ContainerDied","Data":"13fe0f27614d0c26837f83bb6e0f727336fd2761e5c4f72f9ac2682d945328c2"} Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.653590 4837 scope.go:117] "RemoveContainer" containerID="05a9fe216ab76787903f88d032dc165a8ec22819a3f02939ea40e285fdbb5850" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.653737 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-nfc9g" Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.692195 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.697871 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-nfc9g"] Oct 01 07:24:08 crc kubenswrapper[4837]: I1001 07:24:08.944030 4837 scope.go:117] "RemoveContainer" containerID="7607a88a63c494b3dc13e2e72ff3705ae72892ea40bd984c45d3543fb5957bf0" Oct 01 07:24:09 crc kubenswrapper[4837]: I1001 07:24:09.669409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7f2490fe-759d-4263-ae63-0f69c4f78c7d","Type":"ContainerStarted","Data":"afc5fac4c4bdbcdfcb537ec4258dde192ccbd7e0b4e6a075117de4ead1c2405d"} Oct 01 07:24:09 crc kubenswrapper[4837]: I1001 07:24:09.669922 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 01 07:24:09 crc kubenswrapper[4837]: I1001 07:24:09.676076 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerStarted","Data":"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500"} Oct 01 07:24:09 crc kubenswrapper[4837]: I1001 07:24:09.720605 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.784027972 podStartE2EDuration="20.720583659s" podCreationTimestamp="2025-10-01 07:23:49 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.453878237 +0000 UTC m=+1097.295485692" lastFinishedPulling="2025-10-01 07:24:07.390433894 +0000 UTC m=+1104.232041379" observedRunningTime="2025-10-01 07:24:09.692443848 +0000 UTC m=+1106.534051323" watchObservedRunningTime="2025-10-01 07:24:09.720583659 +0000 UTC m=+1106.562191114" Oct 01 07:24:09 crc kubenswrapper[4837]: I1001 07:24:09.824707 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" path="/var/lib/kubelet/pods/aaf9c3c5-dc37-415d-bed9-33af89fffd07/volumes" Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.688486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr" event={"ID":"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62","Type":"ContainerStarted","Data":"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.689179 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-kt6jr" Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.692507 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerStarted","Data":"8d7fbe2134351cb2359c79efd5278c7c6f0897a13622bea6a551c459a08e0614"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.695307 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerStarted","Data":"b2dd01f2720f0c29fc64551b59dc951b915dcc75366896c51b0ec3db97f15550"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.697357 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerStarted","Data":"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.700905 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"14422052-034a-4e59-a29b-6be7aa9d077b","Type":"ContainerStarted","Data":"6a5050452166b7757eb54854930552e444559ebd525aa9ed438a4a78f9e85d22"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.701021 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.703034 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerStarted","Data":"101c60691fb7c41b74bbf3f7d9a52d44e7d583f292d0df8ca0ed6767687a2c81"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.705846 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerStarted","Data":"48c9eecb63371cb6c6437405cb59dd74a49403ac0f542fdf6ce9200b6b395b32"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.707846 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerStarted","Data":"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139"} Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.718022 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-kt6jr" podStartSLOduration=8.197625267 podStartE2EDuration="15.718002937s" podCreationTimestamp="2025-10-01 07:23:55 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.723978191 +0000 UTC m=+1097.565585646" lastFinishedPulling="2025-10-01 07:24:08.244355831 +0000 UTC m=+1105.085963316" observedRunningTime="2025-10-01 07:24:10.710040661 +0000 UTC m=+1107.551648136" watchObservedRunningTime="2025-10-01 07:24:10.718002937 +0000 UTC m=+1107.559610412" Oct 01 07:24:10 crc kubenswrapper[4837]: I1001 07:24:10.837701 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.045201409 podStartE2EDuration="19.837667313s" podCreationTimestamp="2025-10-01 07:23:51 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.869983858 +0000 UTC m=+1097.711591313" lastFinishedPulling="2025-10-01 07:24:09.662449762 +0000 UTC m=+1106.504057217" observedRunningTime="2025-10-01 07:24:10.831793609 +0000 UTC m=+1107.673401064" watchObservedRunningTime="2025-10-01 07:24:10.837667313 +0000 UTC m=+1107.679274768" Oct 01 07:24:11 crc kubenswrapper[4837]: I1001 07:24:11.715345 4837 generic.go:334] "Generic (PLEG): container finished" podID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerID="8d7fbe2134351cb2359c79efd5278c7c6f0897a13622bea6a551c459a08e0614" exitCode=0 Oct 01 07:24:11 crc kubenswrapper[4837]: I1001 07:24:11.715412 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerDied","Data":"8d7fbe2134351cb2359c79efd5278c7c6f0897a13622bea6a551c459a08e0614"} Oct 01 07:24:12 crc kubenswrapper[4837]: I1001 07:24:12.725807 4837 generic.go:334] "Generic (PLEG): container finished" podID="44374bf9-c936-4c06-ac74-e80244c09039" containerID="60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500" exitCode=0 Oct 01 07:24:12 crc kubenswrapper[4837]: I1001 07:24:12.726002 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerDied","Data":"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500"} Oct 01 07:24:12 crc kubenswrapper[4837]: I1001 07:24:12.730092 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerStarted","Data":"d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f"} Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.743824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerStarted","Data":"64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900"} Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.744885 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.744920 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.750830 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerStarted","Data":"a33be5ffcfa7d1c7be08f25ec05a88d909c111fd9dde1c9328ba84acd170fc4c"} Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.754638 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerStarted","Data":"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223"} Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.761956 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerStarted","Data":"167c29409de05ee8281b1aed4f4b8f38acfffec5ceb180aebcf6ac61a6a016c5"} Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.781948 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zpx7h" podStartSLOduration=11.55097586 podStartE2EDuration="18.781894098s" podCreationTimestamp="2025-10-01 07:23:55 +0000 UTC" firstStartedPulling="2025-10-01 07:24:01.012330455 +0000 UTC m=+1097.853937910" lastFinishedPulling="2025-10-01 07:24:08.243248683 +0000 UTC m=+1105.084856148" observedRunningTime="2025-10-01 07:24:13.773571344 +0000 UTC m=+1110.615178809" watchObservedRunningTime="2025-10-01 07:24:13.781894098 +0000 UTC m=+1110.623501563" Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.800330 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.4140340160000004 podStartE2EDuration="19.80031116s" podCreationTimestamp="2025-10-01 07:23:54 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.940420973 +0000 UTC m=+1097.782028428" lastFinishedPulling="2025-10-01 07:24:13.326698117 +0000 UTC m=+1110.168305572" observedRunningTime="2025-10-01 07:24:13.793760909 +0000 UTC m=+1110.635368374" watchObservedRunningTime="2025-10-01 07:24:13.80031116 +0000 UTC m=+1110.641918625" Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.818834 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.228625237 podStartE2EDuration="16.818813804s" podCreationTimestamp="2025-10-01 07:23:57 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.737047093 +0000 UTC m=+1097.578654548" lastFinishedPulling="2025-10-01 07:24:13.32723566 +0000 UTC m=+1110.168843115" observedRunningTime="2025-10-01 07:24:13.812859318 +0000 UTC m=+1110.654466773" watchObservedRunningTime="2025-10-01 07:24:13.818813804 +0000 UTC m=+1110.660421269" Oct 01 07:24:13 crc kubenswrapper[4837]: I1001 07:24:13.839260 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.028997529 podStartE2EDuration="26.839231505s" podCreationTimestamp="2025-10-01 07:23:47 +0000 UTC" firstStartedPulling="2025-10-01 07:24:00.579889281 +0000 UTC m=+1097.421496736" lastFinishedPulling="2025-10-01 07:24:07.390123247 +0000 UTC m=+1104.231730712" observedRunningTime="2025-10-01 07:24:13.836186851 +0000 UTC m=+1110.677794326" watchObservedRunningTime="2025-10-01 07:24:13.839231505 +0000 UTC m=+1110.680839000" Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.004710 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.004747 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.072656 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.605016 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.773157 4837 generic.go:334] "Generic (PLEG): container finished" podID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerID="10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea" exitCode=0 Oct 01 07:24:14 crc kubenswrapper[4837]: I1001 07:24:14.773263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerDied","Data":"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea"} Oct 01 07:24:15 crc kubenswrapper[4837]: I1001 07:24:15.464531 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 01 07:24:15 crc kubenswrapper[4837]: I1001 07:24:15.782312 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerStarted","Data":"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e"} Oct 01 07:24:15 crc kubenswrapper[4837]: I1001 07:24:15.815066 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.687295463 podStartE2EDuration="29.815049454s" podCreationTimestamp="2025-10-01 07:23:46 +0000 UTC" firstStartedPulling="2025-10-01 07:23:53.114430166 +0000 UTC m=+1089.956037661" lastFinishedPulling="2025-10-01 07:24:08.242184157 +0000 UTC m=+1105.083791652" observedRunningTime="2025-10-01 07:24:15.810481842 +0000 UTC m=+1112.652089287" watchObservedRunningTime="2025-10-01 07:24:15.815049454 +0000 UTC m=+1112.656656899" Oct 01 07:24:15 crc kubenswrapper[4837]: I1001 07:24:15.841039 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.079286 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77cd7bb9bf-mtq96"] Oct 01 07:24:16 crc kubenswrapper[4837]: E1001 07:24:16.079630 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="dnsmasq-dns" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.079646 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="dnsmasq-dns" Oct 01 07:24:16 crc kubenswrapper[4837]: E1001 07:24:16.079728 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="init" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.079740 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="init" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.079940 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf9c3c5-dc37-415d-bed9-33af89fffd07" containerName="dnsmasq-dns" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.080888 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.088087 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77cd7bb9bf-mtq96"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.088551 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.130887 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.131804 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.133838 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.139208 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264648 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264714 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264758 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264788 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264808 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgzkk\" (UniqueName: \"kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264825 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn97r\" (UniqueName: \"kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264880 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264894 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.264910 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.342937 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cd7bb9bf-mtq96"] Oct 01 07:24:16 crc kubenswrapper[4837]: E1001 07:24:16.343434 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-zn97r ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" podUID="ac36a15b-07cd-442f-98b5-c0feec4d1735" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366365 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366453 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366489 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366540 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366594 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgzkk\" (UniqueName: \"kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366613 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn97r\" (UniqueName: \"kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366668 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366715 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366736 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.366928 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.367656 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.367711 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.367844 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.368000 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.375951 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.389622 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgzkk\" (UniqueName: \"kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.390843 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.392383 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.394144 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn97r\" (UniqueName: \"kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r\") pod \"dnsmasq-dns-77cd7bb9bf-mtq96\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.397094 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.397397 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2vw77\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.408277 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.447152 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.464564 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.514454 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.569234 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.569290 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcg59\" (UniqueName: \"kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.569314 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.569340 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.569406 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.670624 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.670704 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcg59\" (UniqueName: \"kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.670735 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.670761 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.670809 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.671543 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.671581 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.671803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.671891 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.698732 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcg59\" (UniqueName: \"kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59\") pod \"dnsmasq-dns-86ddb7fb65-tfb4p\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.743802 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.788898 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.801545 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.864775 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.886557 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.982291 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb\") pod \"ac36a15b-07cd-442f-98b5-c0feec4d1735\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.982385 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn97r\" (UniqueName: \"kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r\") pod \"ac36a15b-07cd-442f-98b5-c0feec4d1735\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.982421 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc\") pod \"ac36a15b-07cd-442f-98b5-c0feec4d1735\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.982450 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config\") pod \"ac36a15b-07cd-442f-98b5-c0feec4d1735\" (UID: \"ac36a15b-07cd-442f-98b5-c0feec4d1735\") " Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.982810 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac36a15b-07cd-442f-98b5-c0feec4d1735" (UID: "ac36a15b-07cd-442f-98b5-c0feec4d1735"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.983036 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.983129 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac36a15b-07cd-442f-98b5-c0feec4d1735" (UID: "ac36a15b-07cd-442f-98b5-c0feec4d1735"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.984107 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config" (OuterVolumeSpecName: "config") pod "ac36a15b-07cd-442f-98b5-c0feec4d1735" (UID: "ac36a15b-07cd-442f-98b5-c0feec4d1735"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:16 crc kubenswrapper[4837]: I1001 07:24:16.988922 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r" (OuterVolumeSpecName: "kube-api-access-zn97r") pod "ac36a15b-07cd-442f-98b5-c0feec4d1735" (UID: "ac36a15b-07cd-442f-98b5-c0feec4d1735"). InnerVolumeSpecName "kube-api-access-zn97r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.032725 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.039741 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.059142 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.059205 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bdfbk" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.059238 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.059922 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.060097 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.082571 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.084753 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn97r\" (UniqueName: \"kubernetes.io/projected/ac36a15b-07cd-442f-98b5-c0feec4d1735-kube-api-access-zn97r\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.085028 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.085124 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac36a15b-07cd-442f-98b5-c0feec4d1735-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.186721 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.186756 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.186785 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.186816 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.186916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpzrw\" (UniqueName: \"kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.187042 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.187177 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288074 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpzrw\" (UniqueName: \"kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288111 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288213 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288240 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.288726 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.289593 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.289827 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.293380 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.293920 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.295191 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.302899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpzrw\" (UniqueName: \"kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw\") pod \"ovn-northd-0\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.375622 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.795955 4837 generic.go:334] "Generic (PLEG): container finished" podID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerID="c0eefaff1b420ae4a1c96b3aeb67603ed36b3c9ecd70481297dfbf7c87570460" exitCode=0 Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.796013 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" event={"ID":"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67","Type":"ContainerDied","Data":"c0eefaff1b420ae4a1c96b3aeb67603ed36b3c9ecd70481297dfbf7c87570460"} Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.796037 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" event={"ID":"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67","Type":"ContainerStarted","Data":"9756eb7103127d7c6eaa0832b458431226f8a06e2170be3305275f45be41f020"} Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.797405 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2vw77" event={"ID":"28035a28-28d4-46b3-b269-97a523cf731e","Type":"ContainerStarted","Data":"fa33ac993198b8bd4e6712d1dfc94ce0a3b46c83683d716c97d0c41883173646"} Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.797442 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cd7bb9bf-mtq96" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.797450 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2vw77" event={"ID":"28035a28-28d4-46b3-b269-97a523cf731e","Type":"ContainerStarted","Data":"7d44d925e70e92de59c9cd1cf5d9b55b40f013297d40d8cd8e72b2d5b659418d"} Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.853505 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-2vw77" podStartSLOduration=1.853485349 podStartE2EDuration="1.853485349s" podCreationTimestamp="2025-10-01 07:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:24:17.853464239 +0000 UTC m=+1114.695071694" watchObservedRunningTime="2025-10-01 07:24:17.853485349 +0000 UTC m=+1114.695092804" Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.875783 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:24:17 crc kubenswrapper[4837]: W1001 07:24:17.881482 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4140f175_0421_48d7_84e1_11fd6cdcea52.slice/crio-673e7ff209cad99253c12a17b7d404dfd9f4f5912b7718e5f90531dd3bf3ef2a WatchSource:0}: Error finding container 673e7ff209cad99253c12a17b7d404dfd9f4f5912b7718e5f90531dd3bf3ef2a: Status 404 returned error can't find the container with id 673e7ff209cad99253c12a17b7d404dfd9f4f5912b7718e5f90531dd3bf3ef2a Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.916616 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cd7bb9bf-mtq96"] Oct 01 07:24:17 crc kubenswrapper[4837]: I1001 07:24:17.923190 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77cd7bb9bf-mtq96"] Oct 01 07:24:18 crc kubenswrapper[4837]: I1001 07:24:18.272744 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 01 07:24:18 crc kubenswrapper[4837]: I1001 07:24:18.272805 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 01 07:24:18 crc kubenswrapper[4837]: I1001 07:24:18.809563 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerStarted","Data":"673e7ff209cad99253c12a17b7d404dfd9f4f5912b7718e5f90531dd3bf3ef2a"} Oct 01 07:24:18 crc kubenswrapper[4837]: I1001 07:24:18.813484 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" event={"ID":"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67","Type":"ContainerStarted","Data":"af518aa79be249c5dd079020e8fd4ce5155844a7807fbfb5824990405200f7d4"} Oct 01 07:24:18 crc kubenswrapper[4837]: I1001 07:24:18.849430 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" podStartSLOduration=2.84940219 podStartE2EDuration="2.84940219s" podCreationTimestamp="2025-10-01 07:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:24:18.840597804 +0000 UTC m=+1115.682205269" watchObservedRunningTime="2025-10-01 07:24:18.84940219 +0000 UTC m=+1115.691009685" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.310607 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.310651 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.388747 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.827911 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac36a15b-07cd-442f-98b5-c0feec4d1735" path="/var/lib/kubelet/pods/ac36a15b-07cd-442f-98b5-c0feec4d1735/volumes" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.829334 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:19 crc kubenswrapper[4837]: I1001 07:24:19.885142 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.452783 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.522921 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.558282 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.560934 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.566392 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.678517 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlg79\" (UniqueName: \"kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.678578 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.678601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.678628 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.678701 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.780211 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlg79\" (UniqueName: \"kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.780294 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.780332 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.780391 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.780460 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.781366 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.781386 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.781916 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.782085 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.805153 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlg79\" (UniqueName: \"kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79\") pod \"dnsmasq-dns-768c887765-c8gd6\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.845856 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="dnsmasq-dns" containerID="cri-o://af518aa79be249c5dd079020e8fd4ce5155844a7807fbfb5824990405200f7d4" gracePeriod=10 Oct 01 07:24:21 crc kubenswrapper[4837]: I1001 07:24:21.889147 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.669791 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.678181 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.680635 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2kcvm" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.687982 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.688761 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.689151 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.722065 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.797522 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.797577 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.797605 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d76km\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.797668 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.797708 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.903919 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.904029 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.904077 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d76km\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.904134 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.904170 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.904643 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.909269 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.910581 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: E1001 07:24:22.910592 4837 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 07:24:22 crc kubenswrapper[4837]: E1001 07:24:22.910638 4837 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 07:24:22 crc kubenswrapper[4837]: E1001 07:24:22.910678 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift podName:a27ce0b1-38ff-4a33-be51-5ba651e8e075 nodeName:}" failed. No retries permitted until 2025-10-01 07:24:23.410663647 +0000 UTC m=+1120.252271102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift") pod "swift-storage-0" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075") : configmap "swift-ring-files" not found Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.912551 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tvk89"] Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.922981 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tvk89"] Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.923092 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.926106 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.926298 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.927460 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.937551 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d76km\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:22 crc kubenswrapper[4837]: I1001 07:24:22.957422 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005659 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005735 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn879\" (UniqueName: \"kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005772 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005867 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.005916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.080156 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.080208 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.080245 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.080960 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.081015 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86" gracePeriod=600 Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107397 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107442 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107475 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107530 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107577 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn879\" (UniqueName: \"kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.107607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.108233 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.109259 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.110117 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.111116 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.111383 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.111789 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.137874 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn879\" (UniqueName: \"kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879\") pod \"swift-ring-rebalance-tvk89\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.315885 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:23 crc kubenswrapper[4837]: I1001 07:24:23.412031 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:23 crc kubenswrapper[4837]: E1001 07:24:23.412271 4837 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 07:24:23 crc kubenswrapper[4837]: E1001 07:24:23.412317 4837 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 07:24:23 crc kubenswrapper[4837]: E1001 07:24:23.412412 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift podName:a27ce0b1-38ff-4a33-be51-5ba651e8e075 nodeName:}" failed. No retries permitted until 2025-10-01 07:24:24.412383989 +0000 UTC m=+1121.253991474 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift") pod "swift-storage-0" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075") : configmap "swift-ring-files" not found Oct 01 07:24:24 crc kubenswrapper[4837]: I1001 07:24:24.429505 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:24 crc kubenswrapper[4837]: E1001 07:24:24.429669 4837 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 07:24:24 crc kubenswrapper[4837]: E1001 07:24:24.429856 4837 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 07:24:24 crc kubenswrapper[4837]: E1001 07:24:24.429899 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift podName:a27ce0b1-38ff-4a33-be51-5ba651e8e075 nodeName:}" failed. No retries permitted until 2025-10-01 07:24:26.42988247 +0000 UTC m=+1123.271489925 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift") pod "swift-storage-0" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075") : configmap "swift-ring-files" not found Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.903424 4837 generic.go:334] "Generic (PLEG): container finished" podID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerID="af518aa79be249c5dd079020e8fd4ce5155844a7807fbfb5824990405200f7d4" exitCode=0 Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.904071 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" event={"ID":"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67","Type":"ContainerDied","Data":"af518aa79be249c5dd079020e8fd4ce5155844a7807fbfb5824990405200f7d4"} Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.904107 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" event={"ID":"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67","Type":"ContainerDied","Data":"9756eb7103127d7c6eaa0832b458431226f8a06e2170be3305275f45be41f020"} Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.904123 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9756eb7103127d7c6eaa0832b458431226f8a06e2170be3305275f45be41f020" Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.911451 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.912290 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86" exitCode=0 Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.912341 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86"} Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.912367 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92"} Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.912522 4837 scope.go:117] "RemoveContainer" containerID="f223ab176c749626c6638b01d9092ec25a8be3416fa5bbd2b35c5f53be1429eb" Oct 01 07:24:25 crc kubenswrapper[4837]: I1001 07:24:25.925333 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerStarted","Data":"648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23"} Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.031620 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tvk89"] Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.036616 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:24:26 crc kubenswrapper[4837]: W1001 07:24:26.046491 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod851c014f_3aca_4d28_9061_bf7950283ed5.slice/crio-374f3a2f03b81dfeae4b5fc516aacaae085ca0dfc008ed656a5ee391517a61a7 WatchSource:0}: Error finding container 374f3a2f03b81dfeae4b5fc516aacaae085ca0dfc008ed656a5ee391517a61a7: Status 404 returned error can't find the container with id 374f3a2f03b81dfeae4b5fc516aacaae085ca0dfc008ed656a5ee391517a61a7 Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.064103 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb\") pod \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.064220 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcg59\" (UniqueName: \"kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59\") pod \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.064258 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc\") pod \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.064287 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb\") pod \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.064308 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config\") pod \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\" (UID: \"4592fff5-82a9-4f37-ab0e-dbae4eb8cd67\") " Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.069442 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59" (OuterVolumeSpecName: "kube-api-access-zcg59") pod "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" (UID: "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67"). InnerVolumeSpecName "kube-api-access-zcg59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.099772 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" (UID: "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.102912 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config" (OuterVolumeSpecName: "config") pod "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" (UID: "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.107929 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" (UID: "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.110991 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" (UID: "4592fff5-82a9-4f37-ab0e-dbae4eb8cd67"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.169155 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.169192 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcg59\" (UniqueName: \"kubernetes.io/projected/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-kube-api-access-zcg59\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.169201 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.169209 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.169217 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.475862 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:26 crc kubenswrapper[4837]: E1001 07:24:26.476055 4837 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 07:24:26 crc kubenswrapper[4837]: E1001 07:24:26.476097 4837 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 07:24:26 crc kubenswrapper[4837]: E1001 07:24:26.476177 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift podName:a27ce0b1-38ff-4a33-be51-5ba651e8e075 nodeName:}" failed. No retries permitted until 2025-10-01 07:24:30.476153257 +0000 UTC m=+1127.317760742 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift") pod "swift-storage-0" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075") : configmap "swift-ring-files" not found Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.945812 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerStarted","Data":"a699f84318a8fa1ea08dd47672464162bcd065e0a5b23d594adac792bb355f6b"} Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.946630 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.947706 4837 generic.go:334] "Generic (PLEG): container finished" podID="851c014f-3aca-4d28-9061-bf7950283ed5" containerID="9d9b2827a8b5389385e1a05dae5a6da58994705037f70f7fc0f85336cafaa393" exitCode=0 Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.947765 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-768c887765-c8gd6" event={"ID":"851c014f-3aca-4d28-9061-bf7950283ed5","Type":"ContainerDied","Data":"9d9b2827a8b5389385e1a05dae5a6da58994705037f70f7fc0f85336cafaa393"} Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.947785 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-768c887765-c8gd6" event={"ID":"851c014f-3aca-4d28-9061-bf7950283ed5","Type":"ContainerStarted","Data":"374f3a2f03b81dfeae4b5fc516aacaae085ca0dfc008ed656a5ee391517a61a7"} Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.951020 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86ddb7fb65-tfb4p" Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.951352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tvk89" event={"ID":"cacbfd69-af34-4267-8e75-eee0aba34fe9","Type":"ContainerStarted","Data":"cd5848ee47f8591cbd96f38351bb1d623a6fa30b0966b2e97bfcc6a674a2616a"} Oct 01 07:24:26 crc kubenswrapper[4837]: I1001 07:24:26.988460 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.30413974 podStartE2EDuration="10.988434579s" podCreationTimestamp="2025-10-01 07:24:16 +0000 UTC" firstStartedPulling="2025-10-01 07:24:17.888283004 +0000 UTC m=+1114.729890459" lastFinishedPulling="2025-10-01 07:24:25.572577843 +0000 UTC m=+1122.414185298" observedRunningTime="2025-10-01 07:24:26.969878944 +0000 UTC m=+1123.811486439" watchObservedRunningTime="2025-10-01 07:24:26.988434579 +0000 UTC m=+1123.830042044" Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.124906 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.133894 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86ddb7fb65-tfb4p"] Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.550915 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.625548 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.836322 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" path="/var/lib/kubelet/pods/4592fff5-82a9-4f37-ab0e-dbae4eb8cd67/volumes" Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.962488 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-768c887765-c8gd6" event={"ID":"851c014f-3aca-4d28-9061-bf7950283ed5","Type":"ContainerStarted","Data":"e4dd77ad97727c6b68dd69d696c0eb2142ffcf0a7571c82baada0a76a5d39051"} Oct 01 07:24:27 crc kubenswrapper[4837]: I1001 07:24:27.979285 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-768c887765-c8gd6" podStartSLOduration=6.979270496 podStartE2EDuration="6.979270496s" podCreationTimestamp="2025-10-01 07:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:24:27.975732059 +0000 UTC m=+1124.817339514" watchObservedRunningTime="2025-10-01 07:24:27.979270496 +0000 UTC m=+1124.820877951" Oct 01 07:24:28 crc kubenswrapper[4837]: I1001 07:24:28.972183 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.330823 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-fnmv2"] Oct 01 07:24:29 crc kubenswrapper[4837]: E1001 07:24:29.331659 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="init" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.331682 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="init" Oct 01 07:24:29 crc kubenswrapper[4837]: E1001 07:24:29.331751 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="dnsmasq-dns" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.331766 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="dnsmasq-dns" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.332078 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4592fff5-82a9-4f37-ab0e-dbae4eb8cd67" containerName="dnsmasq-dns" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.333007 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.339128 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fnmv2"] Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.435585 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hdxc\" (UniqueName: \"kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc\") pod \"keystone-db-create-fnmv2\" (UID: \"0416426e-d7dc-458e-aa3a-e1c4a160757d\") " pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.537926 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hdxc\" (UniqueName: \"kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc\") pod \"keystone-db-create-fnmv2\" (UID: \"0416426e-d7dc-458e-aa3a-e1c4a160757d\") " pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.553384 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-njhjj"] Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.555178 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-njhjj" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.560563 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-njhjj"] Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.569956 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hdxc\" (UniqueName: \"kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc\") pod \"keystone-db-create-fnmv2\" (UID: \"0416426e-d7dc-458e-aa3a-e1c4a160757d\") " pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.691259 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.741752 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7q2j\" (UniqueName: \"kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j\") pod \"placement-db-create-njhjj\" (UID: \"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7\") " pod="openstack/placement-db-create-njhjj" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.843781 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7q2j\" (UniqueName: \"kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j\") pod \"placement-db-create-njhjj\" (UID: \"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7\") " pod="openstack/placement-db-create-njhjj" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.867609 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7q2j\" (UniqueName: \"kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j\") pod \"placement-db-create-njhjj\" (UID: \"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7\") " pod="openstack/placement-db-create-njhjj" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.928992 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-njhjj" Oct 01 07:24:29 crc kubenswrapper[4837]: I1001 07:24:29.986571 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tvk89" event={"ID":"cacbfd69-af34-4267-8e75-eee0aba34fe9","Type":"ContainerStarted","Data":"6d9db3522d4f0e2faad737251daa28edf3802cc6ddbfe74acf2694f2acb1b8ac"} Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.015990 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tvk89" podStartSLOduration=4.82750136 podStartE2EDuration="8.015964018s" podCreationTimestamp="2025-10-01 07:24:22 +0000 UTC" firstStartedPulling="2025-10-01 07:24:26.039957713 +0000 UTC m=+1122.881565168" lastFinishedPulling="2025-10-01 07:24:29.228420331 +0000 UTC m=+1126.070027826" observedRunningTime="2025-10-01 07:24:30.002728293 +0000 UTC m=+1126.844335748" watchObservedRunningTime="2025-10-01 07:24:30.015964018 +0000 UTC m=+1126.857571503" Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.204742 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fnmv2"] Oct 01 07:24:30 crc kubenswrapper[4837]: W1001 07:24:30.378805 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50a67cce_6fe6_4bf8_9f10_36eb0544e8b7.slice/crio-5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75 WatchSource:0}: Error finding container 5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75: Status 404 returned error can't find the container with id 5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75 Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.379449 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-njhjj"] Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.556612 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:30 crc kubenswrapper[4837]: E1001 07:24:30.556809 4837 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 07:24:30 crc kubenswrapper[4837]: E1001 07:24:30.556844 4837 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 07:24:30 crc kubenswrapper[4837]: E1001 07:24:30.556904 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift podName:a27ce0b1-38ff-4a33-be51-5ba651e8e075 nodeName:}" failed. No retries permitted until 2025-10-01 07:24:38.556883013 +0000 UTC m=+1135.398490478 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift") pod "swift-storage-0" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075") : configmap "swift-ring-files" not found Oct 01 07:24:30 crc kubenswrapper[4837]: E1001 07:24:30.647704 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0416426e_d7dc_458e_aa3a_e1c4a160757d.slice/crio-conmon-81bc85eef741d4dce8bac1d030483f50daf2acf2f5c76f67467319810a845025.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0416426e_d7dc_458e_aa3a_e1c4a160757d.slice/crio-81bc85eef741d4dce8bac1d030483f50daf2acf2f5c76f67467319810a845025.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.995168 4837 generic.go:334] "Generic (PLEG): container finished" podID="0416426e-d7dc-458e-aa3a-e1c4a160757d" containerID="81bc85eef741d4dce8bac1d030483f50daf2acf2f5c76f67467319810a845025" exitCode=0 Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.995232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fnmv2" event={"ID":"0416426e-d7dc-458e-aa3a-e1c4a160757d","Type":"ContainerDied","Data":"81bc85eef741d4dce8bac1d030483f50daf2acf2f5c76f67467319810a845025"} Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.995291 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fnmv2" event={"ID":"0416426e-d7dc-458e-aa3a-e1c4a160757d","Type":"ContainerStarted","Data":"33f83b4e2ea0642135d3a28f46adf3b0455af6ce48fb200e864fb89d5848d95a"} Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.997279 4837 generic.go:334] "Generic (PLEG): container finished" podID="50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" containerID="3aa66d386d064c0c3f0f19d91744583e91dad1847715857b333c84b42876325a" exitCode=0 Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.997420 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-njhjj" event={"ID":"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7","Type":"ContainerDied","Data":"3aa66d386d064c0c3f0f19d91744583e91dad1847715857b333c84b42876325a"} Oct 01 07:24:30 crc kubenswrapper[4837]: I1001 07:24:30.998106 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-njhjj" event={"ID":"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7","Type":"ContainerStarted","Data":"5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75"} Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.520539 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.525361 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-njhjj" Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.704815 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7q2j\" (UniqueName: \"kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j\") pod \"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7\" (UID: \"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7\") " Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.705008 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hdxc\" (UniqueName: \"kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc\") pod \"0416426e-d7dc-458e-aa3a-e1c4a160757d\" (UID: \"0416426e-d7dc-458e-aa3a-e1c4a160757d\") " Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.710111 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc" (OuterVolumeSpecName: "kube-api-access-7hdxc") pod "0416426e-d7dc-458e-aa3a-e1c4a160757d" (UID: "0416426e-d7dc-458e-aa3a-e1c4a160757d"). InnerVolumeSpecName "kube-api-access-7hdxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.725807 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j" (OuterVolumeSpecName: "kube-api-access-n7q2j") pod "50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" (UID: "50a67cce-6fe6-4bf8-9f10-36eb0544e8b7"). InnerVolumeSpecName "kube-api-access-n7q2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.807227 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7q2j\" (UniqueName: \"kubernetes.io/projected/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7-kube-api-access-n7q2j\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:32 crc kubenswrapper[4837]: I1001 07:24:32.807261 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hdxc\" (UniqueName: \"kubernetes.io/projected/0416426e-d7dc-458e-aa3a-e1c4a160757d-kube-api-access-7hdxc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.015988 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fnmv2" event={"ID":"0416426e-d7dc-458e-aa3a-e1c4a160757d","Type":"ContainerDied","Data":"33f83b4e2ea0642135d3a28f46adf3b0455af6ce48fb200e864fb89d5848d95a"} Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.016019 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fnmv2" Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.016027 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f83b4e2ea0642135d3a28f46adf3b0455af6ce48fb200e864fb89d5848d95a" Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.017882 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-njhjj" event={"ID":"50a67cce-6fe6-4bf8-9f10-36eb0544e8b7","Type":"ContainerDied","Data":"5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75"} Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.017942 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bc6fbcb3b5ee824d6fad91460700aa56d4db4e55d43c7194a608a53323ecf75" Oct 01 07:24:33 crc kubenswrapper[4837]: I1001 07:24:33.017962 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-njhjj" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.807864 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-mdbmd"] Oct 01 07:24:34 crc kubenswrapper[4837]: E1001 07:24:34.808830 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0416426e-d7dc-458e-aa3a-e1c4a160757d" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.808856 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0416426e-d7dc-458e-aa3a-e1c4a160757d" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: E1001 07:24:34.808910 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.808923 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.809249 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0416426e-d7dc-458e-aa3a-e1c4a160757d" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.809278 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" containerName="mariadb-database-create" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.810202 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.817944 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mdbmd"] Oct 01 07:24:34 crc kubenswrapper[4837]: I1001 07:24:34.948745 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bd8q\" (UniqueName: \"kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q\") pod \"glance-db-create-mdbmd\" (UID: \"cdf665e4-7264-4c3b-875e-84ef7aeedd96\") " pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:35 crc kubenswrapper[4837]: I1001 07:24:35.051081 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bd8q\" (UniqueName: \"kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q\") pod \"glance-db-create-mdbmd\" (UID: \"cdf665e4-7264-4c3b-875e-84ef7aeedd96\") " pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:35 crc kubenswrapper[4837]: I1001 07:24:35.074520 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bd8q\" (UniqueName: \"kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q\") pod \"glance-db-create-mdbmd\" (UID: \"cdf665e4-7264-4c3b-875e-84ef7aeedd96\") " pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:35 crc kubenswrapper[4837]: I1001 07:24:35.152401 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:35 crc kubenswrapper[4837]: I1001 07:24:35.418569 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mdbmd"] Oct 01 07:24:35 crc kubenswrapper[4837]: W1001 07:24:35.433626 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf665e4_7264_4c3b_875e_84ef7aeedd96.slice/crio-d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002 WatchSource:0}: Error finding container d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002: Status 404 returned error can't find the container with id d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002 Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.058101 4837 generic.go:334] "Generic (PLEG): container finished" podID="cdf665e4-7264-4c3b-875e-84ef7aeedd96" containerID="81870eeb531bc2fdf0c9e9a1cb9ae49dea463978c323ba7711a80e39cc7b2d2e" exitCode=0 Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.058312 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mdbmd" event={"ID":"cdf665e4-7264-4c3b-875e-84ef7aeedd96","Type":"ContainerDied","Data":"81870eeb531bc2fdf0c9e9a1cb9ae49dea463978c323ba7711a80e39cc7b2d2e"} Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.058503 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mdbmd" event={"ID":"cdf665e4-7264-4c3b-875e-84ef7aeedd96","Type":"ContainerStarted","Data":"d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002"} Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.060914 4837 generic.go:334] "Generic (PLEG): container finished" podID="cacbfd69-af34-4267-8e75-eee0aba34fe9" containerID="6d9db3522d4f0e2faad737251daa28edf3802cc6ddbfe74acf2694f2acb1b8ac" exitCode=0 Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.060972 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tvk89" event={"ID":"cacbfd69-af34-4267-8e75-eee0aba34fe9","Type":"ContainerDied","Data":"6d9db3522d4f0e2faad737251daa28edf3802cc6ddbfe74acf2694f2acb1b8ac"} Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.891003 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.981402 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:24:36 crc kubenswrapper[4837]: I1001 07:24:36.981887 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="dnsmasq-dns" containerID="cri-o://cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728" gracePeriod=10 Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.452462 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.516452 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.526048 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.536434 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598722 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598802 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn879\" (UniqueName: \"kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598859 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598901 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598922 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.598951 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.599008 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf\") pod \"cacbfd69-af34-4267-8e75-eee0aba34fe9\" (UID: \"cacbfd69-af34-4267-8e75-eee0aba34fe9\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.599059 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bd8q\" (UniqueName: \"kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q\") pod \"cdf665e4-7264-4c3b-875e-84ef7aeedd96\" (UID: \"cdf665e4-7264-4c3b-875e-84ef7aeedd96\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.600914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.601277 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.604735 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q" (OuterVolumeSpecName: "kube-api-access-9bd8q") pod "cdf665e4-7264-4c3b-875e-84ef7aeedd96" (UID: "cdf665e4-7264-4c3b-875e-84ef7aeedd96"). InnerVolumeSpecName "kube-api-access-9bd8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.604798 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879" (OuterVolumeSpecName: "kube-api-access-gn879") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "kube-api-access-gn879". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.607849 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.621455 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.623939 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts" (OuterVolumeSpecName: "scripts") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.628056 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cacbfd69-af34-4267-8e75-eee0aba34fe9" (UID: "cacbfd69-af34-4267-8e75-eee0aba34fe9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.700759 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bj6\" (UniqueName: \"kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6\") pod \"3a215cd7-0557-42f0-96b0-8ea079026f6b\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.700877 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config\") pod \"3a215cd7-0557-42f0-96b0-8ea079026f6b\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701013 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc\") pod \"3a215cd7-0557-42f0-96b0-8ea079026f6b\" (UID: \"3a215cd7-0557-42f0-96b0-8ea079026f6b\") " Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701310 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701327 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn879\" (UniqueName: \"kubernetes.io/projected/cacbfd69-af34-4267-8e75-eee0aba34fe9-kube-api-access-gn879\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701338 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701347 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701355 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cacbfd69-af34-4267-8e75-eee0aba34fe9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701363 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cacbfd69-af34-4267-8e75-eee0aba34fe9-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701371 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cacbfd69-af34-4267-8e75-eee0aba34fe9-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.701380 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bd8q\" (UniqueName: \"kubernetes.io/projected/cdf665e4-7264-4c3b-875e-84ef7aeedd96-kube-api-access-9bd8q\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.705547 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6" (OuterVolumeSpecName: "kube-api-access-97bj6") pod "3a215cd7-0557-42f0-96b0-8ea079026f6b" (UID: "3a215cd7-0557-42f0-96b0-8ea079026f6b"). InnerVolumeSpecName "kube-api-access-97bj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.734514 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config" (OuterVolumeSpecName: "config") pod "3a215cd7-0557-42f0-96b0-8ea079026f6b" (UID: "3a215cd7-0557-42f0-96b0-8ea079026f6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.748216 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3a215cd7-0557-42f0-96b0-8ea079026f6b" (UID: "3a215cd7-0557-42f0-96b0-8ea079026f6b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.803285 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bj6\" (UniqueName: \"kubernetes.io/projected/3a215cd7-0557-42f0-96b0-8ea079026f6b-kube-api-access-97bj6\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.803345 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:37 crc kubenswrapper[4837]: I1001 07:24:37.803376 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a215cd7-0557-42f0-96b0-8ea079026f6b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.087070 4837 generic.go:334] "Generic (PLEG): container finished" podID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerID="cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728" exitCode=0 Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.087170 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.087208 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" event={"ID":"3a215cd7-0557-42f0-96b0-8ea079026f6b","Type":"ContainerDied","Data":"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728"} Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.087291 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-hcbzv" event={"ID":"3a215cd7-0557-42f0-96b0-8ea079026f6b","Type":"ContainerDied","Data":"a479f71972b0ac246c9858d14e26af52009b872521a5ef05568f002bdfe68306"} Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.087322 4837 scope.go:117] "RemoveContainer" containerID="cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.091475 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mdbmd" event={"ID":"cdf665e4-7264-4c3b-875e-84ef7aeedd96","Type":"ContainerDied","Data":"d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002"} Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.091562 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d202f174c07a0dad22340993638c87711ff566bbbcd384645c77ae4aa1216002" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.091799 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mdbmd" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.100059 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tvk89" event={"ID":"cacbfd69-af34-4267-8e75-eee0aba34fe9","Type":"ContainerDied","Data":"cd5848ee47f8591cbd96f38351bb1d623a6fa30b0966b2e97bfcc6a674a2616a"} Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.100187 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd5848ee47f8591cbd96f38351bb1d623a6fa30b0966b2e97bfcc6a674a2616a" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.100114 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tvk89" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.188666 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.192440 4837 scope.go:117] "RemoveContainer" containerID="8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.196234 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-hcbzv"] Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.222111 4837 scope.go:117] "RemoveContainer" containerID="cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728" Oct 01 07:24:38 crc kubenswrapper[4837]: E1001 07:24:38.222786 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728\": container with ID starting with cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728 not found: ID does not exist" containerID="cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.222834 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728"} err="failed to get container status \"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728\": rpc error: code = NotFound desc = could not find container \"cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728\": container with ID starting with cb9f810af2753566a6fe1605765de82da95ce5ae20dd283c64e3f73e795c2728 not found: ID does not exist" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.222864 4837 scope.go:117] "RemoveContainer" containerID="8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39" Oct 01 07:24:38 crc kubenswrapper[4837]: E1001 07:24:38.223315 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39\": container with ID starting with 8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39 not found: ID does not exist" containerID="8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.223341 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39"} err="failed to get container status \"8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39\": rpc error: code = NotFound desc = could not find container \"8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39\": container with ID starting with 8c4ef641f9616778f6119f4199b2d46ca47db2efe0b14d76309d064c44b7bb39 not found: ID does not exist" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.617550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.626678 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"swift-storage-0\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " pod="openstack/swift-storage-0" Oct 01 07:24:38 crc kubenswrapper[4837]: I1001 07:24:38.906878 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.319005 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-53a7-account-create-gc7dc"] Oct 01 07:24:39 crc kubenswrapper[4837]: E1001 07:24:39.320003 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacbfd69-af34-4267-8e75-eee0aba34fe9" containerName="swift-ring-rebalance" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320044 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacbfd69-af34-4267-8e75-eee0aba34fe9" containerName="swift-ring-rebalance" Oct 01 07:24:39 crc kubenswrapper[4837]: E1001 07:24:39.320067 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="init" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320079 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="init" Oct 01 07:24:39 crc kubenswrapper[4837]: E1001 07:24:39.320102 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf665e4-7264-4c3b-875e-84ef7aeedd96" containerName="mariadb-database-create" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320117 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf665e4-7264-4c3b-875e-84ef7aeedd96" containerName="mariadb-database-create" Oct 01 07:24:39 crc kubenswrapper[4837]: E1001 07:24:39.320144 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="dnsmasq-dns" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320158 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="dnsmasq-dns" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320455 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" containerName="dnsmasq-dns" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320511 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacbfd69-af34-4267-8e75-eee0aba34fe9" containerName="swift-ring-rebalance" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.320529 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf665e4-7264-4c3b-875e-84ef7aeedd96" containerName="mariadb-database-create" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.321478 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.328760 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.345420 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-53a7-account-create-gc7dc"] Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.432186 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-989db\" (UniqueName: \"kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db\") pod \"keystone-53a7-account-create-gc7dc\" (UID: \"f64b4342-3448-432d-88e7-7c182524aa19\") " pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.488804 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.534299 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-989db\" (UniqueName: \"kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db\") pod \"keystone-53a7-account-create-gc7dc\" (UID: \"f64b4342-3448-432d-88e7-7c182524aa19\") " pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.561767 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-989db\" (UniqueName: \"kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db\") pod \"keystone-53a7-account-create-gc7dc\" (UID: \"f64b4342-3448-432d-88e7-7c182524aa19\") " pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.619306 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d99-account-create-grnjj"] Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.620603 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.625180 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.627177 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d99-account-create-grnjj"] Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.661441 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.737453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcskv\" (UniqueName: \"kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv\") pod \"placement-7d99-account-create-grnjj\" (UID: \"c23ccb73-ce30-474a-9939-1adc6bde5d67\") " pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.828273 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a215cd7-0557-42f0-96b0-8ea079026f6b" path="/var/lib/kubelet/pods/3a215cd7-0557-42f0-96b0-8ea079026f6b/volumes" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.839845 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcskv\" (UniqueName: \"kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv\") pod \"placement-7d99-account-create-grnjj\" (UID: \"c23ccb73-ce30-474a-9939-1adc6bde5d67\") " pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.865892 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcskv\" (UniqueName: \"kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv\") pod \"placement-7d99-account-create-grnjj\" (UID: \"c23ccb73-ce30-474a-9939-1adc6bde5d67\") " pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.905898 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-53a7-account-create-gc7dc"] Oct 01 07:24:39 crc kubenswrapper[4837]: W1001 07:24:39.907614 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf64b4342_3448_432d_88e7_7c182524aa19.slice/crio-afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35 WatchSource:0}: Error finding container afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35: Status 404 returned error can't find the container with id afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35 Oct 01 07:24:39 crc kubenswrapper[4837]: I1001 07:24:39.942614 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.127897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53a7-account-create-gc7dc" event={"ID":"f64b4342-3448-432d-88e7-7c182524aa19","Type":"ContainerStarted","Data":"e975bd72ab746ac62c6e87b203ec843569239bcdf63e46adc77064a72d2c8af5"} Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.128247 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53a7-account-create-gc7dc" event={"ID":"f64b4342-3448-432d-88e7-7c182524aa19","Type":"ContainerStarted","Data":"afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35"} Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.131816 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"9afbec9311de400244922af066bd3bc1dac95bb0d3f80c0b8798a6f61fca1cae"} Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.153165 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-53a7-account-create-gc7dc" podStartSLOduration=1.153144216 podStartE2EDuration="1.153144216s" podCreationTimestamp="2025-10-01 07:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:24:40.142970866 +0000 UTC m=+1136.984578331" watchObservedRunningTime="2025-10-01 07:24:40.153144216 +0000 UTC m=+1136.994751681" Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.424252 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d99-account-create-grnjj"] Oct 01 07:24:40 crc kubenswrapper[4837]: W1001 07:24:40.430601 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc23ccb73_ce30_474a_9939_1adc6bde5d67.slice/crio-42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9 WatchSource:0}: Error finding container 42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9: Status 404 returned error can't find the container with id 42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9 Oct 01 07:24:40 crc kubenswrapper[4837]: I1001 07:24:40.623836 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kt6jr" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" probeResult="failure" output=< Oct 01 07:24:40 crc kubenswrapper[4837]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 07:24:40 crc kubenswrapper[4837]: > Oct 01 07:24:40 crc kubenswrapper[4837]: E1001 07:24:40.910062 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc23ccb73_ce30_474a_9939_1adc6bde5d67.slice/crio-conmon-3604553f9371c79bc1695100ccdedabfe3a3e93d224aaa7de52af1ec682e6662.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.146775 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"6d2afe280ee355f22575011affdf80ede81294951985ed4fdfaf0a0a8bd04577"} Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.149148 4837 generic.go:334] "Generic (PLEG): container finished" podID="f64b4342-3448-432d-88e7-7c182524aa19" containerID="e975bd72ab746ac62c6e87b203ec843569239bcdf63e46adc77064a72d2c8af5" exitCode=0 Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.149207 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53a7-account-create-gc7dc" event={"ID":"f64b4342-3448-432d-88e7-7c182524aa19","Type":"ContainerDied","Data":"e975bd72ab746ac62c6e87b203ec843569239bcdf63e46adc77064a72d2c8af5"} Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.152087 4837 generic.go:334] "Generic (PLEG): container finished" podID="c23ccb73-ce30-474a-9939-1adc6bde5d67" containerID="3604553f9371c79bc1695100ccdedabfe3a3e93d224aaa7de52af1ec682e6662" exitCode=0 Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.152156 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d99-account-create-grnjj" event={"ID":"c23ccb73-ce30-474a-9939-1adc6bde5d67","Type":"ContainerDied","Data":"3604553f9371c79bc1695100ccdedabfe3a3e93d224aaa7de52af1ec682e6662"} Oct 01 07:24:41 crc kubenswrapper[4837]: I1001 07:24:41.152197 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d99-account-create-grnjj" event={"ID":"c23ccb73-ce30-474a-9939-1adc6bde5d67","Type":"ContainerStarted","Data":"42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9"} Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.166654 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"2204c01a3c6692934f14277b4d5cbf6cca7a5c501820bb3fa3389ae18fea6f76"} Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.167099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"8caed28b92cf49fa4a285edcda33f9590fa05d856cd61e8bcce897cc1725a303"} Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.167118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"c67bf25539192ca5553d8c61e702cf3adae4424f5c481f14196a5218a1c14dde"} Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.169889 4837 generic.go:334] "Generic (PLEG): container finished" podID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerID="69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139" exitCode=0 Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.169984 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerDied","Data":"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139"} Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.802575 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.835995 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.911198 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcskv\" (UniqueName: \"kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv\") pod \"c23ccb73-ce30-474a-9939-1adc6bde5d67\" (UID: \"c23ccb73-ce30-474a-9939-1adc6bde5d67\") " Oct 01 07:24:42 crc kubenswrapper[4837]: I1001 07:24:42.919004 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv" (OuterVolumeSpecName: "kube-api-access-tcskv") pod "c23ccb73-ce30-474a-9939-1adc6bde5d67" (UID: "c23ccb73-ce30-474a-9939-1adc6bde5d67"). InnerVolumeSpecName "kube-api-access-tcskv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.012645 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-989db\" (UniqueName: \"kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db\") pod \"f64b4342-3448-432d-88e7-7c182524aa19\" (UID: \"f64b4342-3448-432d-88e7-7c182524aa19\") " Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.013256 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcskv\" (UniqueName: \"kubernetes.io/projected/c23ccb73-ce30-474a-9939-1adc6bde5d67-kube-api-access-tcskv\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.018804 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db" (OuterVolumeSpecName: "kube-api-access-989db") pod "f64b4342-3448-432d-88e7-7c182524aa19" (UID: "f64b4342-3448-432d-88e7-7c182524aa19"). InnerVolumeSpecName "kube-api-access-989db". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.115011 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-989db\" (UniqueName: \"kubernetes.io/projected/f64b4342-3448-432d-88e7-7c182524aa19-kube-api-access-989db\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.186558 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"feb371a16a28f5d9590fbbaf472765ac149570124d961c2b6244deaac26355e5"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.186717 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"9290d642faa7786d6b897ad12a0754d4d61afe7b651852d4dab287495dddc287"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.189257 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-53a7-account-create-gc7dc" event={"ID":"f64b4342-3448-432d-88e7-7c182524aa19","Type":"ContainerDied","Data":"afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.189302 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afa01c7970bab657c88a2993e344367dc411b64e490395f1ba9440944a5e2b35" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.189364 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-53a7-account-create-gc7dc" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.190637 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d99-account-create-grnjj" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.190670 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d99-account-create-grnjj" event={"ID":"c23ccb73-ce30-474a-9939-1adc6bde5d67","Type":"ContainerDied","Data":"42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.190722 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42068fdd54d576a43fd05b742e398d9b5178c293f022efa926502156dc7fd3e9" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.192449 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerStarted","Data":"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.192836 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.193668 4837 generic.go:334] "Generic (PLEG): container finished" podID="96e5feff-74ce-434a-8606-85c51844deb9" containerID="101c60691fb7c41b74bbf3f7d9a52d44e7d583f292d0df8ca0ed6767687a2c81" exitCode=0 Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.193725 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerDied","Data":"101c60691fb7c41b74bbf3f7d9a52d44e7d583f292d0df8ca0ed6767687a2c81"} Oct 01 07:24:43 crc kubenswrapper[4837]: I1001 07:24:43.224590 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.086011179 podStartE2EDuration="59.224566021s" podCreationTimestamp="2025-10-01 07:23:44 +0000 UTC" firstStartedPulling="2025-10-01 07:23:59.970233521 +0000 UTC m=+1096.811840986" lastFinishedPulling="2025-10-01 07:24:07.108788333 +0000 UTC m=+1103.950395828" observedRunningTime="2025-10-01 07:24:43.217529759 +0000 UTC m=+1140.059137244" watchObservedRunningTime="2025-10-01 07:24:43.224566021 +0000 UTC m=+1140.066173516" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.207649 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerStarted","Data":"3f3ea1fd9b8bc426d4733c4c8e9a603415784a810c094ebe41057d96f5b75c2e"} Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.208024 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.214425 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"b5626cefa6e75a0a68d30120dccd151dc40d2c9a0f69f966c98d8c027ef04148"} Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.214472 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"dbb2c5bf1f5499683d8d50b2b442d96eef02c58748d98bc5851c45b9a40621df"} Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.249580 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.747257483 podStartE2EDuration="59.249565436s" podCreationTimestamp="2025-10-01 07:23:45 +0000 UTC" firstStartedPulling="2025-10-01 07:23:58.88683842 +0000 UTC m=+1095.728445915" lastFinishedPulling="2025-10-01 07:24:07.389146403 +0000 UTC m=+1104.230753868" observedRunningTime="2025-10-01 07:24:44.242779209 +0000 UTC m=+1141.084386664" watchObservedRunningTime="2025-10-01 07:24:44.249565436 +0000 UTC m=+1141.091172891" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.942976 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6315-account-create-rb6zg"] Oct 01 07:24:44 crc kubenswrapper[4837]: E1001 07:24:44.943520 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23ccb73-ce30-474a-9939-1adc6bde5d67" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.943539 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23ccb73-ce30-474a-9939-1adc6bde5d67" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: E1001 07:24:44.943568 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64b4342-3448-432d-88e7-7c182524aa19" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.943574 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64b4342-3448-432d-88e7-7c182524aa19" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.943748 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f64b4342-3448-432d-88e7-7c182524aa19" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.943762 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23ccb73-ce30-474a-9939-1adc6bde5d67" containerName="mariadb-account-create" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.944225 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.945794 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 01 07:24:44 crc kubenswrapper[4837]: I1001 07:24:44.958025 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6315-account-create-rb6zg"] Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.045620 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s88l\" (UniqueName: \"kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l\") pod \"glance-6315-account-create-rb6zg\" (UID: \"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970\") " pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.146961 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s88l\" (UniqueName: \"kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l\") pod \"glance-6315-account-create-rb6zg\" (UID: \"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970\") " pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.172417 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s88l\" (UniqueName: \"kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l\") pod \"glance-6315-account-create-rb6zg\" (UID: \"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970\") " pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.226866 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"fd642928ddcba69b2aa30d3fdad7ae984f220ec2171638957d4a6f7678a20b45"} Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.227578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"06276e79f10f7789a399e46f0f9ba7ceccd3ca7926331f03e09147648c3966d4"} Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.227670 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"80729d7b80fe7ca0a76deb60aded53df6ae5bafa464acedacb113c4fcfa46c02"} Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.258745 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.557372 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6315-account-create-rb6zg"] Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.585515 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kt6jr" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" probeResult="failure" output=< Oct 01 07:24:45 crc kubenswrapper[4837]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 07:24:45 crc kubenswrapper[4837]: > Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.683228 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.692522 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.944496 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kt6jr-config-x2gff"] Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.947094 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.948953 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 07:24:45 crc kubenswrapper[4837]: I1001 07:24:45.956727 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr-config-x2gff"] Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060139 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060224 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060253 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060269 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060306 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.060323 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82pg8\" (UniqueName: \"kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161156 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161243 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161265 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161280 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161310 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161326 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82pg8\" (UniqueName: \"kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.161857 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.162124 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.162197 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.162212 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.163353 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.181536 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82pg8\" (UniqueName: \"kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8\") pod \"ovn-controller-kt6jr-config-x2gff\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.234580 4837 generic.go:334] "Generic (PLEG): container finished" podID="b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" containerID="a622535f33e24a5a3d6940ec66cb10d593f27ec6d5ccecd06dc67b1ffadadcf0" exitCode=0 Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.234673 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6315-account-create-rb6zg" event={"ID":"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970","Type":"ContainerDied","Data":"a622535f33e24a5a3d6940ec66cb10d593f27ec6d5ccecd06dc67b1ffadadcf0"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.236219 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6315-account-create-rb6zg" event={"ID":"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970","Type":"ContainerStarted","Data":"eb4981ae9c808d96e810314582e679e623f33fd703d8e921b216dcc23e28607c"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.243003 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"4eb16d26f4116a6a74e9d0b4accc3991ac0da36440e5760b73b34784878eb561"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.243037 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"0bfa759f888d0212de7ee4b0712ad5d33add96370a9084085bce043b2d375d44"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.243048 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"2067fe40fd9f9d75412077801cde319e8068868f481690bf6f201f36f8f71642"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.243057 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerStarted","Data":"708ad62022e8e219ec2f8ff1ce8f2f1e128a8e14443aec6458580876f7636171"} Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.264677 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.304789 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.135305059 podStartE2EDuration="25.304766063s" podCreationTimestamp="2025-10-01 07:24:21 +0000 UTC" firstStartedPulling="2025-10-01 07:24:39.49038164 +0000 UTC m=+1136.331989095" lastFinishedPulling="2025-10-01 07:24:44.659842634 +0000 UTC m=+1141.501450099" observedRunningTime="2025-10-01 07:24:46.281367018 +0000 UTC m=+1143.122974513" watchObservedRunningTime="2025-10-01 07:24:46.304766063 +0000 UTC m=+1143.146373548" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.591161 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr-config-x2gff"] Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.652503 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.654251 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.668627 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.681334 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773091 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773133 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773328 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773443 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996k5\" (UniqueName: \"kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773467 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.773490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.874946 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996k5\" (UniqueName: \"kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.874992 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.875020 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.875078 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.875099 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.875149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.876029 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.876765 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.877270 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.877789 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.878342 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.904862 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996k5\" (UniqueName: \"kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5\") pod \"dnsmasq-dns-674cd4cb49-vrtnq\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:46 crc kubenswrapper[4837]: I1001 07:24:46.978946 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.259896 4837 generic.go:334] "Generic (PLEG): container finished" podID="4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" containerID="a98f366483806d112bcd2392d65b5f235d68f2b5b319dc8a8571c949a57bb438" exitCode=0 Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.262260 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-x2gff" event={"ID":"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5","Type":"ContainerDied","Data":"a98f366483806d112bcd2392d65b5f235d68f2b5b319dc8a8571c949a57bb438"} Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.262294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-x2gff" event={"ID":"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5","Type":"ContainerStarted","Data":"ef54454541eac6666cbeaa668fbad87c2825617e47a872eb9237525fe4ead4cf"} Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.424586 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.508833 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.587230 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s88l\" (UniqueName: \"kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l\") pod \"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970\" (UID: \"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970\") " Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.591649 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l" (OuterVolumeSpecName: "kube-api-access-8s88l") pod "b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" (UID: "b3c4895a-6bd7-4bd6-9beb-bb61bd04b970"). InnerVolumeSpecName "kube-api-access-8s88l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:47 crc kubenswrapper[4837]: I1001 07:24:47.688800 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s88l\" (UniqueName: \"kubernetes.io/projected/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970-kube-api-access-8s88l\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.274235 4837 generic.go:334] "Generic (PLEG): container finished" podID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerID="0ce5c4756fd7fd10bb8c6fed6b1f61429d949908eaa333aecb37a612cb0bbc21" exitCode=0 Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.274345 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" event={"ID":"17d2153c-d0ae-481f-8c8c-29c602e23114","Type":"ContainerDied","Data":"0ce5c4756fd7fd10bb8c6fed6b1f61429d949908eaa333aecb37a612cb0bbc21"} Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.274386 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" event={"ID":"17d2153c-d0ae-481f-8c8c-29c602e23114","Type":"ContainerStarted","Data":"64dead3fa2a04a252430397e51aa28994b9323f8d87efb9918f4c1700dd0e0cb"} Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.279542 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6315-account-create-rb6zg" event={"ID":"b3c4895a-6bd7-4bd6-9beb-bb61bd04b970","Type":"ContainerDied","Data":"eb4981ae9c808d96e810314582e679e623f33fd703d8e921b216dcc23e28607c"} Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.279587 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6315-account-create-rb6zg" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.279601 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb4981ae9c808d96e810314582e679e623f33fd703d8e921b216dcc23e28607c" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.597801 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.713519 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.713901 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82pg8\" (UniqueName: \"kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.713934 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.713999 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714018 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714124 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn\") pod \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\" (UID: \"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5\") " Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714288 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714355 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714443 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run" (OuterVolumeSpecName: "var-run") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714434 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714624 4837 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714646 4837 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714661 4837 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714675 4837 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.714984 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts" (OuterVolumeSpecName: "scripts") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.717712 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8" (OuterVolumeSpecName: "kube-api-access-82pg8") pod "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" (UID: "4d9ab5d2-5045-499b-9bd5-eda8c36a45f5"). InnerVolumeSpecName "kube-api-access-82pg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.815932 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82pg8\" (UniqueName: \"kubernetes.io/projected/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-kube-api-access-82pg8\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:48 crc kubenswrapper[4837]: I1001 07:24:48.815968 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.292498 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" event={"ID":"17d2153c-d0ae-481f-8c8c-29c602e23114","Type":"ContainerStarted","Data":"ec4455314111452568585c736a7b869b222f7ee6602cde92ca750a4103de523f"} Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.292743 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.295028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-x2gff" event={"ID":"4d9ab5d2-5045-499b-9bd5-eda8c36a45f5","Type":"ContainerDied","Data":"ef54454541eac6666cbeaa668fbad87c2825617e47a872eb9237525fe4ead4cf"} Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.295081 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef54454541eac6666cbeaa668fbad87c2825617e47a872eb9237525fe4ead4cf" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.295085 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-x2gff" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.322208 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" podStartSLOduration=3.322187034 podStartE2EDuration="3.322187034s" podCreationTimestamp="2025-10-01 07:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:24:49.315481549 +0000 UTC m=+1146.157089044" watchObservedRunningTime="2025-10-01 07:24:49.322187034 +0000 UTC m=+1146.163794489" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.737898 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kt6jr-config-x2gff"] Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.755409 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kt6jr-config-x2gff"] Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.832137 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" path="/var/lib/kubelet/pods/4d9ab5d2-5045-499b-9bd5-eda8c36a45f5/volumes" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.833062 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kt6jr-config-67lph"] Oct 01 07:24:49 crc kubenswrapper[4837]: E1001 07:24:49.833512 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" containerName="mariadb-account-create" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.833544 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" containerName="mariadb-account-create" Oct 01 07:24:49 crc kubenswrapper[4837]: E1001 07:24:49.833597 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" containerName="ovn-config" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.833612 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" containerName="ovn-config" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.833945 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9ab5d2-5045-499b-9bd5-eda8c36a45f5" containerName="ovn-config" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.833990 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" containerName="mariadb-account-create" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.835201 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.837956 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.845971 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr-config-67lph"] Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.934790 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.935093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.935341 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.935432 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.935594 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6m5\" (UniqueName: \"kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:49 crc kubenswrapper[4837]: I1001 07:24:49.935803 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037299 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037352 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037375 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6m5\" (UniqueName: \"kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037411 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037455 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037475 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037905 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037944 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.037909 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.038232 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.040315 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.061945 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ksdnf"] Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.063091 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.065128 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8mfp2" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.066009 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.071214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6m5\" (UniqueName: \"kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5\") pod \"ovn-controller-kt6jr-config-67lph\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.084232 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ksdnf"] Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.138842 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.139077 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.139249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqj6m\" (UniqueName: \"kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.139384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.156221 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.240936 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqj6m\" (UniqueName: \"kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.240982 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.241053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.241084 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.244340 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.247081 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.255178 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.276063 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqj6m\" (UniqueName: \"kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m\") pod \"glance-db-sync-ksdnf\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.430431 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ksdnf" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.601420 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-kt6jr" Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.632874 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kt6jr-config-67lph"] Oct 01 07:24:50 crc kubenswrapper[4837]: W1001 07:24:50.636177 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f7f641_bf76_4772_ba60_18ec9238f41f.slice/crio-0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866 WatchSource:0}: Error finding container 0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866: Status 404 returned error can't find the container with id 0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866 Oct 01 07:24:50 crc kubenswrapper[4837]: I1001 07:24:50.961427 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ksdnf"] Oct 01 07:24:50 crc kubenswrapper[4837]: W1001 07:24:50.968219 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4a0fc8_f0d7_4bc4_9c3a_d53421e88e88.slice/crio-016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc WatchSource:0}: Error finding container 016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc: Status 404 returned error can't find the container with id 016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc Oct 01 07:24:51 crc kubenswrapper[4837]: E1001 07:24:51.132880 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f7f641_bf76_4772_ba60_18ec9238f41f.slice/crio-conmon-26e6a34c5516afcbbf5945753d417ac28fae4221a6a145124aa496a1fd01b8b4.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:24:51 crc kubenswrapper[4837]: I1001 07:24:51.322147 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ksdnf" event={"ID":"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88","Type":"ContainerStarted","Data":"016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc"} Oct 01 07:24:51 crc kubenswrapper[4837]: I1001 07:24:51.324056 4837 generic.go:334] "Generic (PLEG): container finished" podID="e4f7f641-bf76-4772-ba60-18ec9238f41f" containerID="26e6a34c5516afcbbf5945753d417ac28fae4221a6a145124aa496a1fd01b8b4" exitCode=0 Oct 01 07:24:51 crc kubenswrapper[4837]: I1001 07:24:51.324094 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-67lph" event={"ID":"e4f7f641-bf76-4772-ba60-18ec9238f41f","Type":"ContainerDied","Data":"26e6a34c5516afcbbf5945753d417ac28fae4221a6a145124aa496a1fd01b8b4"} Oct 01 07:24:51 crc kubenswrapper[4837]: I1001 07:24:51.324118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-67lph" event={"ID":"e4f7f641-bf76-4772-ba60-18ec9238f41f","Type":"ContainerStarted","Data":"0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866"} Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.639172 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682287 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682340 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6m5\" (UniqueName: \"kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682487 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682498 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682524 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682563 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn\") pod \"e4f7f641-bf76-4772-ba60-18ec9238f41f\" (UID: \"e4f7f641-bf76-4772-ba60-18ec9238f41f\") " Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.682982 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run" (OuterVolumeSpecName: "var-run") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.683090 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.683047 4837 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.683543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.683991 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts" (OuterVolumeSpecName: "scripts") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.688001 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5" (OuterVolumeSpecName: "kube-api-access-rw6m5") pod "e4f7f641-bf76-4772-ba60-18ec9238f41f" (UID: "e4f7f641-bf76-4772-ba60-18ec9238f41f"). InnerVolumeSpecName "kube-api-access-rw6m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.784155 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.784181 4837 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f7f641-bf76-4772-ba60-18ec9238f41f-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.784193 4837 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.784201 4837 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e4f7f641-bf76-4772-ba60-18ec9238f41f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:52 crc kubenswrapper[4837]: I1001 07:24:52.784230 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6m5\" (UniqueName: \"kubernetes.io/projected/e4f7f641-bf76-4772-ba60-18ec9238f41f-kube-api-access-rw6m5\") on node \"crc\" DevicePath \"\"" Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.344407 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr-config-67lph" event={"ID":"e4f7f641-bf76-4772-ba60-18ec9238f41f","Type":"ContainerDied","Data":"0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866"} Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.344953 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0be41eaccfb1a572ba7bc1d7905f096c57a8c3d707945f9140f777c542d57866" Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.345047 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr-config-67lph" Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.718101 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kt6jr-config-67lph"] Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.724686 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kt6jr-config-67lph"] Oct 01 07:24:53 crc kubenswrapper[4837]: I1001 07:24:53.830137 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f7f641-bf76-4772-ba60-18ec9238f41f" path="/var/lib/kubelet/pods/e4f7f641-bf76-4772-ba60-18ec9238f41f/volumes" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.356930 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.662775 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-kjn5l"] Oct 01 07:24:56 crc kubenswrapper[4837]: E1001 07:24:56.663384 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f7f641-bf76-4772-ba60-18ec9238f41f" containerName="ovn-config" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.663468 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f7f641-bf76-4772-ba60-18ec9238f41f" containerName="ovn-config" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.663723 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f7f641-bf76-4772-ba60-18ec9238f41f" containerName="ovn-config" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.664337 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kjn5l" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.670556 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kjn5l"] Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.701906 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.776671 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-sj7hz"] Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.777755 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sj7hz" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.787398 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sj7hz"] Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.864530 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhnfk\" (UniqueName: \"kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk\") pod \"cinder-db-create-kjn5l\" (UID: \"fc0af960-55d3-436a-b532-f8460076932e\") " pod="openstack/cinder-db-create-kjn5l" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.924906 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-pn9tb"] Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.925891 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.927608 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lcpw5" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.927766 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.928107 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.928165 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.966286 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dk4b\" (UniqueName: \"kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b\") pod \"barbican-db-create-sj7hz\" (UID: \"39965847-66b7-45ad-b24e-1ce81e39515b\") " pod="openstack/barbican-db-create-sj7hz" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.966400 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhnfk\" (UniqueName: \"kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk\") pod \"cinder-db-create-kjn5l\" (UID: \"fc0af960-55d3-436a-b532-f8460076932e\") " pod="openstack/cinder-db-create-kjn5l" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.966492 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-rh7lh"] Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.967440 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rh7lh" Oct 01 07:24:56 crc kubenswrapper[4837]: I1001 07:24:56.981500 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.001061 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-rh7lh"] Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.004686 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pn9tb"] Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.006627 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhnfk\" (UniqueName: \"kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk\") pod \"cinder-db-create-kjn5l\" (UID: \"fc0af960-55d3-436a-b532-f8460076932e\") " pod="openstack/cinder-db-create-kjn5l" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.053864 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.054120 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-768c887765-c8gd6" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="dnsmasq-dns" containerID="cri-o://e4dd77ad97727c6b68dd69d696c0eb2142ffcf0a7571c82baada0a76a5d39051" gracePeriod=10 Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.069220 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6pv8\" (UniqueName: \"kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.069331 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp9dq\" (UniqueName: \"kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq\") pod \"neutron-db-create-rh7lh\" (UID: \"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5\") " pod="openstack/neutron-db-create-rh7lh" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.069396 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dk4b\" (UniqueName: \"kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b\") pod \"barbican-db-create-sj7hz\" (UID: \"39965847-66b7-45ad-b24e-1ce81e39515b\") " pod="openstack/barbican-db-create-sj7hz" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.069453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.069525 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.133844 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dk4b\" (UniqueName: \"kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b\") pod \"barbican-db-create-sj7hz\" (UID: \"39965847-66b7-45ad-b24e-1ce81e39515b\") " pod="openstack/barbican-db-create-sj7hz" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.172835 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp9dq\" (UniqueName: \"kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq\") pod \"neutron-db-create-rh7lh\" (UID: \"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5\") " pod="openstack/neutron-db-create-rh7lh" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.172909 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.172960 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.173015 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6pv8\" (UniqueName: \"kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.177201 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.177601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.196892 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6pv8\" (UniqueName: \"kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8\") pod \"keystone-db-sync-pn9tb\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.197441 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp9dq\" (UniqueName: \"kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq\") pod \"neutron-db-create-rh7lh\" (UID: \"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5\") " pod="openstack/neutron-db-create-rh7lh" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.240119 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.280068 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rh7lh" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.280929 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kjn5l" Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.385543 4837 generic.go:334] "Generic (PLEG): container finished" podID="851c014f-3aca-4d28-9061-bf7950283ed5" containerID="e4dd77ad97727c6b68dd69d696c0eb2142ffcf0a7571c82baada0a76a5d39051" exitCode=0 Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.385597 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-768c887765-c8gd6" event={"ID":"851c014f-3aca-4d28-9061-bf7950283ed5","Type":"ContainerDied","Data":"e4dd77ad97727c6b68dd69d696c0eb2142ffcf0a7571c82baada0a76a5d39051"} Oct 01 07:24:57 crc kubenswrapper[4837]: I1001 07:24:57.400903 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sj7hz" Oct 01 07:25:01 crc kubenswrapper[4837]: I1001 07:25:01.889909 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-768c887765-c8gd6" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.502321 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.597844 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb\") pod \"851c014f-3aca-4d28-9061-bf7950283ed5\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.598155 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config\") pod \"851c014f-3aca-4d28-9061-bf7950283ed5\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.598295 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb\") pod \"851c014f-3aca-4d28-9061-bf7950283ed5\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.598329 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc\") pod \"851c014f-3aca-4d28-9061-bf7950283ed5\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.598351 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlg79\" (UniqueName: \"kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79\") pod \"851c014f-3aca-4d28-9061-bf7950283ed5\" (UID: \"851c014f-3aca-4d28-9061-bf7950283ed5\") " Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.605326 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79" (OuterVolumeSpecName: "kube-api-access-dlg79") pod "851c014f-3aca-4d28-9061-bf7950283ed5" (UID: "851c014f-3aca-4d28-9061-bf7950283ed5"). InnerVolumeSpecName "kube-api-access-dlg79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.643360 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config" (OuterVolumeSpecName: "config") pod "851c014f-3aca-4d28-9061-bf7950283ed5" (UID: "851c014f-3aca-4d28-9061-bf7950283ed5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.645068 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "851c014f-3aca-4d28-9061-bf7950283ed5" (UID: "851c014f-3aca-4d28-9061-bf7950283ed5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.650597 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "851c014f-3aca-4d28-9061-bf7950283ed5" (UID: "851c014f-3aca-4d28-9061-bf7950283ed5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.651949 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "851c014f-3aca-4d28-9061-bf7950283ed5" (UID: "851c014f-3aca-4d28-9061-bf7950283ed5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.700309 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.700338 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.700348 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.700357 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/851c014f-3aca-4d28-9061-bf7950283ed5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.700365 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlg79\" (UniqueName: \"kubernetes.io/projected/851c014f-3aca-4d28-9061-bf7950283ed5-kube-api-access-dlg79\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.779507 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sj7hz"] Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.786487 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pn9tb"] Oct 01 07:25:04 crc kubenswrapper[4837]: W1001 07:25:04.787773 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39965847_66b7_45ad_b24e_1ce81e39515b.slice/crio-2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73 WatchSource:0}: Error finding container 2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73: Status 404 returned error can't find the container with id 2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73 Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.872180 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-rh7lh"] Oct 01 07:25:04 crc kubenswrapper[4837]: I1001 07:25:04.878871 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kjn5l"] Oct 01 07:25:04 crc kubenswrapper[4837]: W1001 07:25:04.885138 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0af960_55d3_436a_b532_f8460076932e.slice/crio-b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565 WatchSource:0}: Error finding container b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565: Status 404 returned error can't find the container with id b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565 Oct 01 07:25:04 crc kubenswrapper[4837]: W1001 07:25:04.886968 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c35a796_0dbb_41de_b8c3_222c4dcd1ef5.slice/crio-55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89 WatchSource:0}: Error finding container 55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89: Status 404 returned error can't find the container with id 55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89 Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.520810 4837 generic.go:334] "Generic (PLEG): container finished" podID="39965847-66b7-45ad-b24e-1ce81e39515b" containerID="2f1b15f1d32699a3f4010b989a02711f339c01163b9b29fe3a66969c2d8cdcff" exitCode=0 Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.521087 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sj7hz" event={"ID":"39965847-66b7-45ad-b24e-1ce81e39515b","Type":"ContainerDied","Data":"2f1b15f1d32699a3f4010b989a02711f339c01163b9b29fe3a66969c2d8cdcff"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.522844 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sj7hz" event={"ID":"39965847-66b7-45ad-b24e-1ce81e39515b","Type":"ContainerStarted","Data":"2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.532317 4837 generic.go:334] "Generic (PLEG): container finished" podID="7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" containerID="19c640a121f67ee824cde93f790747e72586eaf1ece8a807acdd3a97e2cbe3ac" exitCode=0 Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.532497 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rh7lh" event={"ID":"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5","Type":"ContainerDied","Data":"19c640a121f67ee824cde93f790747e72586eaf1ece8a807acdd3a97e2cbe3ac"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.533012 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rh7lh" event={"ID":"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5","Type":"ContainerStarted","Data":"55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.545270 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pn9tb" event={"ID":"85661570-740d-45f3-a34c-bc575512ca64","Type":"ContainerStarted","Data":"cfd548ff8d54f92ed911298d7a5f9714c2da7062e7f124107828895218eadce8"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.558025 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-768c887765-c8gd6" Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.558035 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-768c887765-c8gd6" event={"ID":"851c014f-3aca-4d28-9061-bf7950283ed5","Type":"ContainerDied","Data":"374f3a2f03b81dfeae4b5fc516aacaae085ca0dfc008ed656a5ee391517a61a7"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.558169 4837 scope.go:117] "RemoveContainer" containerID="e4dd77ad97727c6b68dd69d696c0eb2142ffcf0a7571c82baada0a76a5d39051" Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.566259 4837 generic.go:334] "Generic (PLEG): container finished" podID="fc0af960-55d3-436a-b532-f8460076932e" containerID="f79c5da489b3e4e2a44ecd79261d14e1290a38a0e382e01f266b6136c16b1250" exitCode=0 Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.566349 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kjn5l" event={"ID":"fc0af960-55d3-436a-b532-f8460076932e","Type":"ContainerDied","Data":"f79c5da489b3e4e2a44ecd79261d14e1290a38a0e382e01f266b6136c16b1250"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.566384 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kjn5l" event={"ID":"fc0af960-55d3-436a-b532-f8460076932e","Type":"ContainerStarted","Data":"b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.569965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ksdnf" event={"ID":"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88","Type":"ContainerStarted","Data":"3a08d66e194419e08949044103a9a74620bbcd529385fbab247b8a2c78c247bd"} Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.610174 4837 scope.go:117] "RemoveContainer" containerID="9d9b2827a8b5389385e1a05dae5a6da58994705037f70f7fc0f85336cafaa393" Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.618624 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ksdnf" podStartSLOduration=2.240739378 podStartE2EDuration="15.618606935s" podCreationTimestamp="2025-10-01 07:24:50 +0000 UTC" firstStartedPulling="2025-10-01 07:24:50.970757172 +0000 UTC m=+1147.812364637" lastFinishedPulling="2025-10-01 07:25:04.348624739 +0000 UTC m=+1161.190232194" observedRunningTime="2025-10-01 07:25:05.60534434 +0000 UTC m=+1162.446951805" watchObservedRunningTime="2025-10-01 07:25:05.618606935 +0000 UTC m=+1162.460214390" Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.627596 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.633578 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-768c887765-c8gd6"] Oct 01 07:25:05 crc kubenswrapper[4837]: I1001 07:25:05.834834 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" path="/var/lib/kubelet/pods/851c014f-3aca-4d28-9061-bf7950283ed5/volumes" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.127534 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rh7lh" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.133584 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kjn5l" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.141771 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sj7hz" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.209073 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp9dq\" (UniqueName: \"kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq\") pod \"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5\" (UID: \"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5\") " Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.213886 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq" (OuterVolumeSpecName: "kube-api-access-mp9dq") pod "7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" (UID: "7c35a796-0dbb-41de-b8c3-222c4dcd1ef5"). InnerVolumeSpecName "kube-api-access-mp9dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.310293 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dk4b\" (UniqueName: \"kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b\") pod \"39965847-66b7-45ad-b24e-1ce81e39515b\" (UID: \"39965847-66b7-45ad-b24e-1ce81e39515b\") " Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.310646 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhnfk\" (UniqueName: \"kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk\") pod \"fc0af960-55d3-436a-b532-f8460076932e\" (UID: \"fc0af960-55d3-436a-b532-f8460076932e\") " Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.311156 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp9dq\" (UniqueName: \"kubernetes.io/projected/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5-kube-api-access-mp9dq\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.313576 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk" (OuterVolumeSpecName: "kube-api-access-hhnfk") pod "fc0af960-55d3-436a-b532-f8460076932e" (UID: "fc0af960-55d3-436a-b532-f8460076932e"). InnerVolumeSpecName "kube-api-access-hhnfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.313745 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b" (OuterVolumeSpecName: "kube-api-access-9dk4b") pod "39965847-66b7-45ad-b24e-1ce81e39515b" (UID: "39965847-66b7-45ad-b24e-1ce81e39515b"). InnerVolumeSpecName "kube-api-access-9dk4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.413075 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dk4b\" (UniqueName: \"kubernetes.io/projected/39965847-66b7-45ad-b24e-1ce81e39515b-kube-api-access-9dk4b\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.413120 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhnfk\" (UniqueName: \"kubernetes.io/projected/fc0af960-55d3-436a-b532-f8460076932e-kube-api-access-hhnfk\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.646983 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pn9tb" event={"ID":"85661570-740d-45f3-a34c-bc575512ca64","Type":"ContainerStarted","Data":"5d7fc883157d22705620cceb566a7750f43040d9aac88d9b6d381c18b7520ea5"} Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.648929 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kjn5l" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.648928 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kjn5l" event={"ID":"fc0af960-55d3-436a-b532-f8460076932e","Type":"ContainerDied","Data":"b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565"} Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.648962 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b039fa9948cbbc189d73bad054b24cac13b3e04dc3fd3d2c0a3b5898b025e565" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.650822 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sj7hz" event={"ID":"39965847-66b7-45ad-b24e-1ce81e39515b","Type":"ContainerDied","Data":"2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73"} Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.650845 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2665a141ef0cffcf6e97bbdf75a929dc4c56b4480de6fe1e95ffa39e9608ad73" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.650846 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sj7hz" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.652135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-rh7lh" event={"ID":"7c35a796-0dbb-41de-b8c3-222c4dcd1ef5","Type":"ContainerDied","Data":"55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89"} Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.652162 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55ff17af2e19ba692ce52b6e7b87ca916d71d29544fe20f22534720d37295e89" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.652202 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-rh7lh" Oct 01 07:25:11 crc kubenswrapper[4837]: I1001 07:25:11.671950 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-pn9tb" podStartSLOduration=9.26424043 podStartE2EDuration="15.670286158s" podCreationTimestamp="2025-10-01 07:24:56 +0000 UTC" firstStartedPulling="2025-10-01 07:25:04.788245837 +0000 UTC m=+1161.629853282" lastFinishedPulling="2025-10-01 07:25:11.194291555 +0000 UTC m=+1168.035899010" observedRunningTime="2025-10-01 07:25:11.661789967 +0000 UTC m=+1168.503397432" watchObservedRunningTime="2025-10-01 07:25:11.670286158 +0000 UTC m=+1168.511893653" Oct 01 07:25:13 crc kubenswrapper[4837]: I1001 07:25:13.683356 4837 generic.go:334] "Generic (PLEG): container finished" podID="7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" containerID="3a08d66e194419e08949044103a9a74620bbcd529385fbab247b8a2c78c247bd" exitCode=0 Oct 01 07:25:13 crc kubenswrapper[4837]: I1001 07:25:13.683493 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ksdnf" event={"ID":"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88","Type":"ContainerDied","Data":"3a08d66e194419e08949044103a9a74620bbcd529385fbab247b8a2c78c247bd"} Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.171969 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ksdnf" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.289610 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data\") pod \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.289677 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqj6m\" (UniqueName: \"kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m\") pod \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.289778 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data\") pod \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.289846 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle\") pod \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\" (UID: \"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88\") " Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.298158 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m" (OuterVolumeSpecName: "kube-api-access-cqj6m") pod "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" (UID: "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88"). InnerVolumeSpecName "kube-api-access-cqj6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.298235 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" (UID: "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.341384 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" (UID: "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.356033 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data" (OuterVolumeSpecName: "config-data") pod "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" (UID: "7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.391817 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.391855 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqj6m\" (UniqueName: \"kubernetes.io/projected/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-kube-api-access-cqj6m\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.391865 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.391874 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.708335 4837 generic.go:334] "Generic (PLEG): container finished" podID="85661570-740d-45f3-a34c-bc575512ca64" containerID="5d7fc883157d22705620cceb566a7750f43040d9aac88d9b6d381c18b7520ea5" exitCode=0 Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.708454 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pn9tb" event={"ID":"85661570-740d-45f3-a34c-bc575512ca64","Type":"ContainerDied","Data":"5d7fc883157d22705620cceb566a7750f43040d9aac88d9b6d381c18b7520ea5"} Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.711884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ksdnf" event={"ID":"7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88","Type":"ContainerDied","Data":"016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc"} Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.711947 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="016dff0f34782b0f32470919a0409a7cde829b25cff8cbf546f9fc0b8b5ff0fc" Oct 01 07:25:15 crc kubenswrapper[4837]: I1001 07:25:15.712014 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ksdnf" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312510 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312830 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0af960-55d3-436a-b532-f8460076932e" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312842 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0af960-55d3-436a-b532-f8460076932e" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312853 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39965847-66b7-45ad-b24e-1ce81e39515b" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312858 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="39965847-66b7-45ad-b24e-1ce81e39515b" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312866 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" containerName="glance-db-sync" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312872 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" containerName="glance-db-sync" Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312884 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="dnsmasq-dns" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312889 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="dnsmasq-dns" Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312901 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="init" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312907 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="init" Oct 01 07:25:16 crc kubenswrapper[4837]: E1001 07:25:16.312913 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.312920 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313054 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" containerName="glance-db-sync" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313068 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="851c014f-3aca-4d28-9061-bf7950283ed5" containerName="dnsmasq-dns" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313080 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="39965847-66b7-45ad-b24e-1ce81e39515b" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313097 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0af960-55d3-436a-b532-f8460076932e" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313108 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" containerName="mariadb-database-create" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.313853 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.331373 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407084 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8qlh\" (UniqueName: \"kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407167 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407190 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407210 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407241 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.407263 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508746 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508809 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508838 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508860 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.508917 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8qlh\" (UniqueName: \"kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.509869 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.510600 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.510749 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.511094 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.511255 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.524990 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8qlh\" (UniqueName: \"kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh\") pod \"dnsmasq-dns-85799f4775-rwd4f\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:16 crc kubenswrapper[4837]: I1001 07:25:16.627901 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.727797 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pn9tb" event={"ID":"85661570-740d-45f3-a34c-bc575512ca64","Type":"ContainerDied","Data":"cfd548ff8d54f92ed911298d7a5f9714c2da7062e7f124107828895218eadce8"} Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.738889 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfd548ff8d54f92ed911298d7a5f9714c2da7062e7f124107828895218eadce8" Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.754621 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.903024 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:17 crc kubenswrapper[4837]: W1001 07:25:17.912654 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddace7edf_8803_4075_b76c_ccbb151ec97c.slice/crio-9639c2d783cc5638616bb8ca1f9dc4a486857f447d26e49accd36cf5a9abe38b WatchSource:0}: Error finding container 9639c2d783cc5638616bb8ca1f9dc4a486857f447d26e49accd36cf5a9abe38b: Status 404 returned error can't find the container with id 9639c2d783cc5638616bb8ca1f9dc4a486857f447d26e49accd36cf5a9abe38b Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.930377 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6pv8\" (UniqueName: \"kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8\") pod \"85661570-740d-45f3-a34c-bc575512ca64\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.930465 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle\") pod \"85661570-740d-45f3-a34c-bc575512ca64\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.930486 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data\") pod \"85661570-740d-45f3-a34c-bc575512ca64\" (UID: \"85661570-740d-45f3-a34c-bc575512ca64\") " Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.933425 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8" (OuterVolumeSpecName: "kube-api-access-q6pv8") pod "85661570-740d-45f3-a34c-bc575512ca64" (UID: "85661570-740d-45f3-a34c-bc575512ca64"). InnerVolumeSpecName "kube-api-access-q6pv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.977543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85661570-740d-45f3-a34c-bc575512ca64" (UID: "85661570-740d-45f3-a34c-bc575512ca64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:17 crc kubenswrapper[4837]: I1001 07:25:17.984759 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data" (OuterVolumeSpecName: "config-data") pod "85661570-740d-45f3-a34c-bc575512ca64" (UID: "85661570-740d-45f3-a34c-bc575512ca64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.031682 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.031728 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85661570-740d-45f3-a34c-bc575512ca64-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.031738 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6pv8\" (UniqueName: \"kubernetes.io/projected/85661570-740d-45f3-a34c-bc575512ca64-kube-api-access-q6pv8\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.739221 4837 generic.go:334] "Generic (PLEG): container finished" podID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerID="9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0" exitCode=0 Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.739263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" event={"ID":"dace7edf-8803-4075-b76c-ccbb151ec97c","Type":"ContainerDied","Data":"9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0"} Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.739297 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pn9tb" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.739322 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" event={"ID":"dace7edf-8803-4075-b76c-ccbb151ec97c","Type":"ContainerStarted","Data":"9639c2d783cc5638616bb8ca1f9dc4a486857f447d26e49accd36cf5a9abe38b"} Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.986294 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-h7xnb"] Oct 01 07:25:18 crc kubenswrapper[4837]: E1001 07:25:18.986910 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85661570-740d-45f3-a34c-bc575512ca64" containerName="keystone-db-sync" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.986930 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="85661570-740d-45f3-a34c-bc575512ca64" containerName="keystone-db-sync" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.987141 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="85661570-740d-45f3-a34c-bc575512ca64" containerName="keystone-db-sync" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.987631 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.989885 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.990070 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.990323 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 07:25:18 crc kubenswrapper[4837]: I1001 07:25:18.990469 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lcpw5" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.002026 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h7xnb"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.041160 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.076632 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7597c59797-rwvvn"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.086631 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.095368 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7597c59797-rwvvn"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150316 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150452 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150495 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjlqc\" (UniqueName: \"kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150515 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.150536 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.189376 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.195282 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.197447 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.202494 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.209863 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.251934 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252076 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252162 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252227 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252337 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252405 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252465 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252530 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252617 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjlqc\" (UniqueName: \"kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252707 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252773 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95jc9\" (UniqueName: \"kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.252836 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.257253 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.257281 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.258454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.258473 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.258893 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.272793 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjlqc\" (UniqueName: \"kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc\") pod \"keystone-bootstrap-h7xnb\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.292374 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7597c59797-rwvvn"] Oct 01 07:25:19 crc kubenswrapper[4837]: E1001 07:25:19.293842 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-95jc9 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7597c59797-rwvvn" podUID="d49fa526-e817-4dea-abe9-9343094e7f2f" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.311088 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.346402 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8fqtj"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.347534 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.351047 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.351217 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.351321 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cm8kg" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354038 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354090 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354120 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354138 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354176 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354215 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jc9\" (UniqueName: \"kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354235 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hds\" (UniqueName: \"kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354253 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354275 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354292 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354312 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354331 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.354362 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.355161 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.355850 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.356361 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.357261 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.357350 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.357459 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.358670 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.373653 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8fqtj"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.385350 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jc9\" (UniqueName: \"kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9\") pod \"dnsmasq-dns-7597c59797-rwvvn\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.414805 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455222 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtzn\" (UniqueName: \"kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455272 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455290 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455315 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455337 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdtbj\" (UniqueName: \"kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455359 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455402 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455420 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455460 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hds\" (UniqueName: \"kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455478 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455497 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455517 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455534 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455554 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.455572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.456912 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.457191 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.457272 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.457362 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.460161 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.460293 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.460557 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.461597 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.474627 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hds\" (UniqueName: \"kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds\") pod \"ceilometer-0\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.550936 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559515 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdtbj\" (UniqueName: \"kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559585 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559603 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559626 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559666 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559710 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559742 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559791 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtzn\" (UniqueName: \"kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559814 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.559833 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.560846 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.564105 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.564382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.564988 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.566393 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.567017 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.567229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.569923 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.570134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.579986 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdtbj\" (UniqueName: \"kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj\") pod \"placement-db-sync-8fqtj\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.584937 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtzn\" (UniqueName: \"kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn\") pod \"dnsmasq-dns-65d7cf9447-t7wtm\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.696217 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.711786 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.780666 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.780734 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" event={"ID":"dace7edf-8803-4075-b76c-ccbb151ec97c","Type":"ContainerStarted","Data":"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963"} Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.780761 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="dnsmasq-dns" containerID="cri-o://441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963" gracePeriod=10 Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.781439 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.796034 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.819073 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" podStartSLOduration=3.819058632 podStartE2EDuration="3.819058632s" podCreationTimestamp="2025-10-01 07:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:19.816901328 +0000 UTC m=+1176.658508783" watchObservedRunningTime="2025-10-01 07:25:19.819058632 +0000 UTC m=+1176.660666087" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969269 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969739 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969773 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969792 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969870 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95jc9\" (UniqueName: \"kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.969962 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb\") pod \"d49fa526-e817-4dea-abe9-9343094e7f2f\" (UID: \"d49fa526-e817-4dea-abe9-9343094e7f2f\") " Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.971564 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config" (OuterVolumeSpecName: "config") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.972006 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.972416 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.972915 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.973859 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:19 crc kubenswrapper[4837]: I1001 07:25:19.976392 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9" (OuterVolumeSpecName: "kube-api-access-95jc9") pod "d49fa526-e817-4dea-abe9-9343094e7f2f" (UID: "d49fa526-e817-4dea-abe9-9343094e7f2f"). InnerVolumeSpecName "kube-api-access-95jc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.002318 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h7xnb"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.002368 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:25:20 crc kubenswrapper[4837]: W1001 07:25:20.006825 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f55b0fc_aa86_4bc0_a072_03f52fa47f87.slice/crio-18c66ecbd2e9e0d825743ad001d1b061d1e67fd3b3598f865300922005a24f50 WatchSource:0}: Error finding container 18c66ecbd2e9e0d825743ad001d1b061d1e67fd3b3598f865300922005a24f50: Status 404 returned error can't find the container with id 18c66ecbd2e9e0d825743ad001d1b061d1e67fd3b3598f865300922005a24f50 Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072676 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072741 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072753 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072770 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072780 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d49fa526-e817-4dea-abe9-9343094e7f2f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.072790 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95jc9\" (UniqueName: \"kubernetes.io/projected/d49fa526-e817-4dea-abe9-9343094e7f2f-kube-api-access-95jc9\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.089217 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.093677 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.095034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.095945 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.096662 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.096861 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.097876 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8mfp2" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.141460 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.142835 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.152568 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.153038 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.159805 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.180992 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8fqtj"] Oct 01 07:25:20 crc kubenswrapper[4837]: W1001 07:25:20.182367 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod201dd153_cadb_48ca_bddc_b7ba0622f39a.slice/crio-0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875 WatchSource:0}: Error finding container 0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875: Status 404 returned error can't find the container with id 0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875 Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275165 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275198 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275218 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275435 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275480 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275510 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdbwr\" (UniqueName: \"kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275666 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275715 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275784 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275813 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qffk\" (UniqueName: \"kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275863 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275882 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.275898 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.276434 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.305668 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380147 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8qlh\" (UniqueName: \"kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380234 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380347 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380432 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380833 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380877 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380908 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380929 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qffk\" (UniqueName: \"kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380951 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380972 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.380993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381080 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381099 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381134 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381201 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381227 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdbwr\" (UniqueName: \"kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.381265 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.385520 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.385956 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.385997 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.388807 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.389327 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.389540 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.394118 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.394289 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.394409 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.394758 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.398574 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.412902 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh" (OuterVolumeSpecName: "kube-api-access-p8qlh") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "kube-api-access-p8qlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.413817 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qffk\" (UniqueName: \"kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.414216 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.414432 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdbwr\" (UniqueName: \"kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.415241 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.416617 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.442641 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.445862 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.468943 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.489306 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config" (OuterVolumeSpecName: "config") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.489829 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.489849 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8qlh\" (UniqueName: \"kubernetes.io/projected/dace7edf-8803-4075-b76c-ccbb151ec97c-kube-api-access-p8qlh\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.532606 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.548195 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: E1001 07:25:20.563011 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb podName:dace7edf-8803-4075-b76c-ccbb151ec97c nodeName:}" failed. No retries permitted until 2025-10-01 07:25:21.062985592 +0000 UTC m=+1177.904593047 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c") : error deleting /var/lib/kubelet/pods/dace7edf-8803-4075-b76c-ccbb151ec97c/volume-subpaths: remove /var/lib/kubelet/pods/dace7edf-8803-4075-b76c-ccbb151ec97c/volume-subpaths: no such file or directory Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.563198 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.599790 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.599993 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.600065 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.721550 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.814194 4837 generic.go:334] "Generic (PLEG): container finished" podID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerID="441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963" exitCode=0 Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.814246 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.814317 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" event={"ID":"dace7edf-8803-4075-b76c-ccbb151ec97c","Type":"ContainerDied","Data":"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.814396 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85799f4775-rwd4f" event={"ID":"dace7edf-8803-4075-b76c-ccbb151ec97c","Type":"ContainerDied","Data":"9639c2d783cc5638616bb8ca1f9dc4a486857f447d26e49accd36cf5a9abe38b"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.814415 4837 scope.go:117] "RemoveContainer" containerID="441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.824021 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" event={"ID":"2ee17db9-cebc-49b6-8fee-7dedb0c04622","Type":"ContainerStarted","Data":"c567c0873991d3d42b8a6e3b05cbf8f2345ebb032c8dc58413999ca772c5bc58"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.830661 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerStarted","Data":"18c66ecbd2e9e0d825743ad001d1b061d1e67fd3b3598f865300922005a24f50"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.832308 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h7xnb" event={"ID":"938d9193-5002-48cc-868c-b0e961325b2b","Type":"ContainerStarted","Data":"0fc6fdf9472d712ace5bbb580bcdc39380a6fc356087908b1642817b1194f4c7"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.832349 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h7xnb" event={"ID":"938d9193-5002-48cc-868c-b0e961325b2b","Type":"ContainerStarted","Data":"b7ee123a24f1ba606e2bedfdb8972d8b07868bbc30dc617ebcb8b7b58354dec3"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.854542 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-h7xnb" podStartSLOduration=2.854528932 podStartE2EDuration="2.854528932s" podCreationTimestamp="2025-10-01 07:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:20.853905587 +0000 UTC m=+1177.695513042" watchObservedRunningTime="2025-10-01 07:25:20.854528932 +0000 UTC m=+1177.696136387" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.885813 4837 scope.go:117] "RemoveContainer" containerID="9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.888361 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7597c59797-rwvvn" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.888395 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8fqtj" event={"ID":"201dd153-cadb-48ca-bddc-b7ba0622f39a","Type":"ContainerStarted","Data":"0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875"} Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.930587 4837 scope.go:117] "RemoveContainer" containerID="441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963" Oct 01 07:25:20 crc kubenswrapper[4837]: E1001 07:25:20.931173 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963\": container with ID starting with 441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963 not found: ID does not exist" containerID="441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.931216 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963"} err="failed to get container status \"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963\": rpc error: code = NotFound desc = could not find container \"441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963\": container with ID starting with 441a04c0009f141be38733eccc6f066414972a472b74dc6094e273e9db645963 not found: ID does not exist" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.931243 4837 scope.go:117] "RemoveContainer" containerID="9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0" Oct 01 07:25:20 crc kubenswrapper[4837]: E1001 07:25:20.931576 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0\": container with ID starting with 9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0 not found: ID does not exist" containerID="9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.931605 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0"} err="failed to get container status \"9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0\": rpc error: code = NotFound desc = could not find container \"9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0\": container with ID starting with 9fe3fb2082dfb1137fa24f9a942df32eab2cb838bc91956b9615b3b1591b62b0 not found: ID does not exist" Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.982891 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7597c59797-rwvvn"] Oct 01 07:25:20 crc kubenswrapper[4837]: I1001 07:25:20.996362 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7597c59797-rwvvn"] Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.121911 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") pod \"dace7edf-8803-4075-b76c-ccbb151ec97c\" (UID: \"dace7edf-8803-4075-b76c-ccbb151ec97c\") " Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.122567 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dace7edf-8803-4075-b76c-ccbb151ec97c" (UID: "dace7edf-8803-4075-b76c-ccbb151ec97c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.193496 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.225296 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dace7edf-8803-4075-b76c-ccbb151ec97c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.462247 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.473200 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85799f4775-rwd4f"] Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.537401 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:21 crc kubenswrapper[4837]: W1001 07:25:21.547941 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8be2274e_c0d9_431c_8be3_d9562ce9687f.slice/crio-042f0f9aa190c0c6bb782ac5d6a8bfc73ac96b9bb5c6e85fd6cecefab85b9739 WatchSource:0}: Error finding container 042f0f9aa190c0c6bb782ac5d6a8bfc73ac96b9bb5c6e85fd6cecefab85b9739: Status 404 returned error can't find the container with id 042f0f9aa190c0c6bb782ac5d6a8bfc73ac96b9bb5c6e85fd6cecefab85b9739 Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.826996 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49fa526-e817-4dea-abe9-9343094e7f2f" path="/var/lib/kubelet/pods/d49fa526-e817-4dea-abe9-9343094e7f2f/volumes" Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.827638 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" path="/var/lib/kubelet/pods/dace7edf-8803-4075-b76c-ccbb151ec97c/volumes" Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.902080 4837 generic.go:334] "Generic (PLEG): container finished" podID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerID="1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0" exitCode=0 Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.902135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" event={"ID":"2ee17db9-cebc-49b6-8fee-7dedb0c04622","Type":"ContainerDied","Data":"1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0"} Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.908001 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerStarted","Data":"8047fca459449bca59406932b56fb9ba7a77a892987a364e2ce919ca94fe4075"} Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.908033 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerStarted","Data":"28fbb842834a0068b7b5561f82c094f0a746093c92005607f76c6a22da3e247c"} Oct 01 07:25:21 crc kubenswrapper[4837]: I1001 07:25:21.910531 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerStarted","Data":"042f0f9aa190c0c6bb782ac5d6a8bfc73ac96b9bb5c6e85fd6cecefab85b9739"} Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.380368 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.455654 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.459913 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.930477 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" event={"ID":"2ee17db9-cebc-49b6-8fee-7dedb0c04622","Type":"ContainerStarted","Data":"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb"} Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.931090 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.933737 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerStarted","Data":"84296860c2fbd41fb9b394e5d352926b38914b373cc7d669d06ee5707c33976b"} Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.933812 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-log" containerID="cri-o://8047fca459449bca59406932b56fb9ba7a77a892987a364e2ce919ca94fe4075" gracePeriod=30 Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.933842 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-httpd" containerID="cri-o://84296860c2fbd41fb9b394e5d352926b38914b373cc7d669d06ee5707c33976b" gracePeriod=30 Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.961952 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerStarted","Data":"2a1288c3f8cce06ab9a18499284a82f7fafb671f721969fd9e8e3dd7a37b51a9"} Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.962004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerStarted","Data":"20f815cded2225ae7c100cbe305cb35ac7f014f71b4c77c3566a91dc76b31402"} Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.962126 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-log" containerID="cri-o://20f815cded2225ae7c100cbe305cb35ac7f014f71b4c77c3566a91dc76b31402" gracePeriod=30 Oct 01 07:25:22 crc kubenswrapper[4837]: I1001 07:25:22.962393 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-httpd" containerID="cri-o://2a1288c3f8cce06ab9a18499284a82f7fafb671f721969fd9e8e3dd7a37b51a9" gracePeriod=30 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.053605 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.053585222 podStartE2EDuration="4.053585222s" podCreationTimestamp="2025-10-01 07:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:23.053162421 +0000 UTC m=+1179.894769876" watchObservedRunningTime="2025-10-01 07:25:23.053585222 +0000 UTC m=+1179.895192677" Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.054761 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" podStartSLOduration=4.05475515 podStartE2EDuration="4.05475515s" podCreationTimestamp="2025-10-01 07:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:23.016249758 +0000 UTC m=+1179.857857213" watchObservedRunningTime="2025-10-01 07:25:23.05475515 +0000 UTC m=+1179.896362605" Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.083996 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.083982233 podStartE2EDuration="4.083982233s" podCreationTimestamp="2025-10-01 07:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:23.082109907 +0000 UTC m=+1179.923717362" watchObservedRunningTime="2025-10-01 07:25:23.083982233 +0000 UTC m=+1179.925589678" Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.972923 4837 generic.go:334] "Generic (PLEG): container finished" podID="938d9193-5002-48cc-868c-b0e961325b2b" containerID="0fc6fdf9472d712ace5bbb580bcdc39380a6fc356087908b1642817b1194f4c7" exitCode=0 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.973275 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h7xnb" event={"ID":"938d9193-5002-48cc-868c-b0e961325b2b","Type":"ContainerDied","Data":"0fc6fdf9472d712ace5bbb580bcdc39380a6fc356087908b1642817b1194f4c7"} Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.976737 4837 generic.go:334] "Generic (PLEG): container finished" podID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerID="2a1288c3f8cce06ab9a18499284a82f7fafb671f721969fd9e8e3dd7a37b51a9" exitCode=143 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.976772 4837 generic.go:334] "Generic (PLEG): container finished" podID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerID="20f815cded2225ae7c100cbe305cb35ac7f014f71b4c77c3566a91dc76b31402" exitCode=143 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.976787 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerDied","Data":"2a1288c3f8cce06ab9a18499284a82f7fafb671f721969fd9e8e3dd7a37b51a9"} Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.976821 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerDied","Data":"20f815cded2225ae7c100cbe305cb35ac7f014f71b4c77c3566a91dc76b31402"} Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.978721 4837 generic.go:334] "Generic (PLEG): container finished" podID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerID="84296860c2fbd41fb9b394e5d352926b38914b373cc7d669d06ee5707c33976b" exitCode=0 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.982807 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerDied","Data":"84296860c2fbd41fb9b394e5d352926b38914b373cc7d669d06ee5707c33976b"} Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.982822 4837 generic.go:334] "Generic (PLEG): container finished" podID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerID="8047fca459449bca59406932b56fb9ba7a77a892987a364e2ce919ca94fe4075" exitCode=143 Oct 01 07:25:23 crc kubenswrapper[4837]: I1001 07:25:23.982855 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerDied","Data":"8047fca459449bca59406932b56fb9ba7a77a892987a364e2ce919ca94fe4075"} Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.655001 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e1a9-account-create-6drcx"] Oct 01 07:25:26 crc kubenswrapper[4837]: E1001 07:25:26.656359 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="dnsmasq-dns" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.656385 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="dnsmasq-dns" Oct 01 07:25:26 crc kubenswrapper[4837]: E1001 07:25:26.656421 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="init" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.656436 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="init" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.656797 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="dace7edf-8803-4075-b76c-ccbb151ec97c" containerName="dnsmasq-dns" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.657626 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.659732 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.668340 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e1a9-account-create-6drcx"] Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.742860 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f4b1-account-create-lzfzs"] Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.744630 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.746831 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcn6k\" (UniqueName: \"kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k\") pod \"cinder-e1a9-account-create-6drcx\" (UID: \"303dfe99-b3ef-4f14-857b-014fd13dce75\") " pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.747369 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.751382 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f4b1-account-create-lzfzs"] Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.848463 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7wfn\" (UniqueName: \"kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn\") pod \"barbican-f4b1-account-create-lzfzs\" (UID: \"d8322b3d-5eac-45bc-ad35-f98e62cf9715\") " pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.848633 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcn6k\" (UniqueName: \"kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k\") pod \"cinder-e1a9-account-create-6drcx\" (UID: \"303dfe99-b3ef-4f14-857b-014fd13dce75\") " pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.869459 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcn6k\" (UniqueName: \"kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k\") pod \"cinder-e1a9-account-create-6drcx\" (UID: \"303dfe99-b3ef-4f14-857b-014fd13dce75\") " pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.943163 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8fcf-account-create-t728v"] Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.944462 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.946218 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.949915 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7wfn\" (UniqueName: \"kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn\") pod \"barbican-f4b1-account-create-lzfzs\" (UID: \"d8322b3d-5eac-45bc-ad35-f98e62cf9715\") " pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.950233 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8fcf-account-create-t728v"] Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.977166 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7wfn\" (UniqueName: \"kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn\") pod \"barbican-f4b1-account-create-lzfzs\" (UID: \"d8322b3d-5eac-45bc-ad35-f98e62cf9715\") " pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:26 crc kubenswrapper[4837]: I1001 07:25:26.986020 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.051546 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbv4s\" (UniqueName: \"kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s\") pod \"neutron-8fcf-account-create-t728v\" (UID: \"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb\") " pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.077045 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.152908 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbv4s\" (UniqueName: \"kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s\") pod \"neutron-8fcf-account-create-t728v\" (UID: \"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb\") " pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.169250 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbv4s\" (UniqueName: \"kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s\") pod \"neutron-8fcf-account-create-t728v\" (UID: \"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb\") " pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.325932 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.920184 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.958738 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968242 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968284 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qffk\" (UniqueName: \"kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968362 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968404 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968437 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968456 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.968564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts\") pod \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\" (UID: \"f6d1430a-9ea2-45a3-aa45-23135466ebe1\") " Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.969116 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs" (OuterVolumeSpecName: "logs") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.979098 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:25:27 crc kubenswrapper[4837]: I1001 07:25:27.995100 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts" (OuterVolumeSpecName: "scripts") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.005274 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.007473 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk" (OuterVolumeSpecName: "kube-api-access-6qffk") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "kube-api-access-6qffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.045457 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f6d1430a-9ea2-45a3-aa45-23135466ebe1","Type":"ContainerDied","Data":"28fbb842834a0068b7b5561f82c094f0a746093c92005607f76c6a22da3e247c"} Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.045512 4837 scope.go:117] "RemoveContainer" containerID="84296860c2fbd41fb9b394e5d352926b38914b373cc7d669d06ee5707c33976b" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.045650 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.055712 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h7xnb" event={"ID":"938d9193-5002-48cc-868c-b0e961325b2b","Type":"ContainerDied","Data":"b7ee123a24f1ba606e2bedfdb8972d8b07868bbc30dc617ebcb8b7b58354dec3"} Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.055746 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7ee123a24f1ba606e2bedfdb8972d8b07868bbc30dc617ebcb8b7b58354dec3" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.055802 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h7xnb" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.070725 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.070861 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.070919 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.070968 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071064 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071103 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjlqc\" (UniqueName: \"kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc\") pod \"938d9193-5002-48cc-868c-b0e961325b2b\" (UID: \"938d9193-5002-48cc-868c-b0e961325b2b\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071415 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071428 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071437 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qffk\" (UniqueName: \"kubernetes.io/projected/f6d1430a-9ea2-45a3-aa45-23135466ebe1-kube-api-access-6qffk\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071446 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d1430a-9ea2-45a3-aa45-23135466ebe1-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.071465 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.078131 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.078361 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc" (OuterVolumeSpecName: "kube-api-access-tjlqc") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "kube-api-access-tjlqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.079675 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.081564 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts" (OuterVolumeSpecName: "scripts") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.086571 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.119686 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.135245 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.145907 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.146092 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data" (OuterVolumeSpecName: "config-data") pod "f6d1430a-9ea2-45a3-aa45-23135466ebe1" (UID: "f6d1430a-9ea2-45a3-aa45-23135466ebe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.161918 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data" (OuterVolumeSpecName: "config-data") pod "938d9193-5002-48cc-868c-b0e961325b2b" (UID: "938d9193-5002-48cc-868c-b0e961325b2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.167839 4837 scope.go:117] "RemoveContainer" containerID="8047fca459449bca59406932b56fb9ba7a77a892987a364e2ce919ca94fe4075" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172579 4837 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172601 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172611 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172621 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172630 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjlqc\" (UniqueName: \"kubernetes.io/projected/938d9193-5002-48cc-868c-b0e961325b2b-kube-api-access-tjlqc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172639 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172647 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172656 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6d1430a-9ea2-45a3-aa45-23135466ebe1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172663 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/938d9193-5002-48cc-868c-b0e961325b2b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.172671 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.325478 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.371247 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375490 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375531 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375633 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375656 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375683 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdbwr\" (UniqueName: \"kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375747 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.375840 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle\") pod \"8be2274e-c0d9-431c-8be3-d9562ce9687f\" (UID: \"8be2274e-c0d9-431c-8be3-d9562ce9687f\") " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.376149 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.376408 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs" (OuterVolumeSpecName: "logs") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.380416 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.396573 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts" (OuterVolumeSpecName: "scripts") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.405275 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr" (OuterVolumeSpecName: "kube-api-access-pdbwr") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "kube-api-access-pdbwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.409832 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.417788 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:28 crc kubenswrapper[4837]: E1001 07:25:28.418644 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418665 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: E1001 07:25:28.418678 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938d9193-5002-48cc-868c-b0e961325b2b" containerName="keystone-bootstrap" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418701 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="938d9193-5002-48cc-868c-b0e961325b2b" containerName="keystone-bootstrap" Oct 01 07:25:28 crc kubenswrapper[4837]: E1001 07:25:28.418720 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418726 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: E1001 07:25:28.418743 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418749 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: E1001 07:25:28.418756 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418762 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418920 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418936 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-httpd" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418943 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418956 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" containerName="glance-log" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.418964 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="938d9193-5002-48cc-868c-b0e961325b2b" containerName="keystone-bootstrap" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.419800 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.422153 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.422344 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.429450 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.438391 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.443082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data" (OuterVolumeSpecName: "config-data") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479696 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62j52\" (UniqueName: \"kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479721 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479745 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479777 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479793 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479811 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479836 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479910 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479921 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479929 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479947 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479955 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8be2274e-c0d9-431c-8be3-d9562ce9687f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479965 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdbwr\" (UniqueName: \"kubernetes.io/projected/8be2274e-c0d9-431c-8be3-d9562ce9687f-kube-api-access-pdbwr\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.479975 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.498370 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f4b1-account-create-lzfzs"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.508840 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8be2274e-c0d9-431c-8be3-d9562ce9687f" (UID: "8be2274e-c0d9-431c-8be3-d9562ce9687f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.518841 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62j52\" (UniqueName: \"kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581528 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581592 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581609 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581627 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581648 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581743 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581790 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.581801 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8be2274e-c0d9-431c-8be3-d9562ce9687f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.582181 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.582623 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.583930 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.592847 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.606976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.612393 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.615823 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.619731 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e1a9-account-create-6drcx"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.625462 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8fcf-account-create-t728v"] Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.630408 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62j52\" (UniqueName: \"kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.674832 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:25:28 crc kubenswrapper[4837]: I1001 07:25:28.753334 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.051994 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-h7xnb"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.057723 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-h7xnb"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.065476 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerStarted","Data":"8fcfd7702a2a1f358988e7055d8a5cfae7b3a6dbbe84448d6f78c2cfbb4d95ad"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.067839 4837 generic.go:334] "Generic (PLEG): container finished" podID="dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" containerID="f0d02bee9b43b1aa4c0d88fee58fc197ff6560dc60daaca70139c6751c5ce0a9" exitCode=0 Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.067936 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8fcf-account-create-t728v" event={"ID":"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb","Type":"ContainerDied","Data":"f0d02bee9b43b1aa4c0d88fee58fc197ff6560dc60daaca70139c6751c5ce0a9"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.067991 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8fcf-account-create-t728v" event={"ID":"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb","Type":"ContainerStarted","Data":"1465846bd5becfd8412919b360cbb9a880ea73f93661b69a764831198c24df2a"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.075493 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8be2274e-c0d9-431c-8be3-d9562ce9687f","Type":"ContainerDied","Data":"042f0f9aa190c0c6bb782ac5d6a8bfc73ac96b9bb5c6e85fd6cecefab85b9739"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.075528 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.075536 4837 scope.go:117] "RemoveContainer" containerID="2a1288c3f8cce06ab9a18499284a82f7fafb671f721969fd9e8e3dd7a37b51a9" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.077975 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8fqtj" event={"ID":"201dd153-cadb-48ca-bddc-b7ba0622f39a","Type":"ContainerStarted","Data":"c3b73bdd602f1754beef682cc6ae2997909ff57156cad987858a8ff3e196d7a8"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.084129 4837 generic.go:334] "Generic (PLEG): container finished" podID="303dfe99-b3ef-4f14-857b-014fd13dce75" containerID="31c1c4531205c7aa377e609c7ac5c5f31a4a0e827612da211359083203b511a5" exitCode=0 Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.084215 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e1a9-account-create-6drcx" event={"ID":"303dfe99-b3ef-4f14-857b-014fd13dce75","Type":"ContainerDied","Data":"31c1c4531205c7aa377e609c7ac5c5f31a4a0e827612da211359083203b511a5"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.084522 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e1a9-account-create-6drcx" event={"ID":"303dfe99-b3ef-4f14-857b-014fd13dce75","Type":"ContainerStarted","Data":"3309e518d42f5f5134132bcf936e06944f0c4061707e0dc96deed769cda635b8"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.086787 4837 generic.go:334] "Generic (PLEG): container finished" podID="d8322b3d-5eac-45bc-ad35-f98e62cf9715" containerID="986d92bb1fcb5055d52a0dfa057171bd964fd3fe435bf9c633eba6301866d672" exitCode=0 Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.086824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f4b1-account-create-lzfzs" event={"ID":"d8322b3d-5eac-45bc-ad35-f98e62cf9715","Type":"ContainerDied","Data":"986d92bb1fcb5055d52a0dfa057171bd964fd3fe435bf9c633eba6301866d672"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.086847 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f4b1-account-create-lzfzs" event={"ID":"d8322b3d-5eac-45bc-ad35-f98e62cf9715","Type":"ContainerStarted","Data":"39127ff4b01518f843247e0843a7458eb77af7672bfa3f2964498f9d04c76fdf"} Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.104430 4837 scope.go:117] "RemoveContainer" containerID="20f815cded2225ae7c100cbe305cb35ac7f014f71b4c77c3566a91dc76b31402" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.110609 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8fqtj" podStartSLOduration=2.337660742 podStartE2EDuration="10.110589941s" podCreationTimestamp="2025-10-01 07:25:19 +0000 UTC" firstStartedPulling="2025-10-01 07:25:20.187226007 +0000 UTC m=+1177.028833462" lastFinishedPulling="2025-10-01 07:25:27.960155196 +0000 UTC m=+1184.801762661" observedRunningTime="2025-10-01 07:25:29.102809758 +0000 UTC m=+1185.944417233" watchObservedRunningTime="2025-10-01 07:25:29.110589941 +0000 UTC m=+1185.952197396" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.162954 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.176737 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.188020 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fpv6z"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.189148 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.192423 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.192533 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.192765 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lcpw5" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.192875 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.197927 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fpv6z"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.210826 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.212222 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.216262 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.216409 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.221611 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303321 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303355 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303398 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303416 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5hjf\" (UniqueName: \"kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303452 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303475 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303493 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303531 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgzhh\" (UniqueName: \"kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303557 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.303604 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.316558 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404724 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404760 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404804 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404824 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404843 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404900 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404927 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404973 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.404990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.405006 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5hjf\" (UniqueName: \"kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.405064 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.405080 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.405118 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.405159 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgzhh\" (UniqueName: \"kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.406865 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.407514 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.409217 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.412592 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.413403 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.413775 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.416277 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.417872 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.417937 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.418098 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.418963 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.426490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgzhh\" (UniqueName: \"kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.442044 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.447468 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5hjf\" (UniqueName: \"kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf\") pod \"keystone-bootstrap-fpv6z\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.453935 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.505193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.535618 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.717905 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.772858 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.773116 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="dnsmasq-dns" containerID="cri-o://ec4455314111452568585c736a7b869b222f7ee6602cde92ca750a4103de523f" gracePeriod=10 Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.835144 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be2274e-c0d9-431c-8be3-d9562ce9687f" path="/var/lib/kubelet/pods/8be2274e-c0d9-431c-8be3-d9562ce9687f/volumes" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.837343 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="938d9193-5002-48cc-868c-b0e961325b2b" path="/var/lib/kubelet/pods/938d9193-5002-48cc-868c-b0e961325b2b/volumes" Oct 01 07:25:29 crc kubenswrapper[4837]: I1001 07:25:29.837980 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d1430a-9ea2-45a3-aa45-23135466ebe1" path="/var/lib/kubelet/pods/f6d1430a-9ea2-45a3-aa45-23135466ebe1/volumes" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.021648 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fpv6z"] Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.105636 4837 generic.go:334] "Generic (PLEG): container finished" podID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerID="ec4455314111452568585c736a7b869b222f7ee6602cde92ca750a4103de523f" exitCode=0 Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.105937 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" event={"ID":"17d2153c-d0ae-481f-8c8c-29c602e23114","Type":"ContainerDied","Data":"ec4455314111452568585c736a7b869b222f7ee6602cde92ca750a4103de523f"} Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.110599 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerStarted","Data":"4f77b3c3d1afee943b4cc1259496e89befab677fcc2ebe167d5561a153e61d31"} Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.222150 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:25:30 crc kubenswrapper[4837]: W1001 07:25:30.454577 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d8e10dd_a7a8_49ce_9fa3_a4a63d02366c.slice/crio-0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa WatchSource:0}: Error finding container 0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa: Status 404 returned error can't find the container with id 0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.776214 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.825166 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.826792 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.845569 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbv4s\" (UniqueName: \"kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s\") pod \"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb\" (UID: \"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.845706 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7wfn\" (UniqueName: \"kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn\") pod \"d8322b3d-5eac-45bc-ad35-f98e62cf9715\" (UID: \"d8322b3d-5eac-45bc-ad35-f98e62cf9715\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.845786 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcn6k\" (UniqueName: \"kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k\") pod \"303dfe99-b3ef-4f14-857b-014fd13dce75\" (UID: \"303dfe99-b3ef-4f14-857b-014fd13dce75\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.851966 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s" (OuterVolumeSpecName: "kube-api-access-sbv4s") pod "dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" (UID: "dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb"). InnerVolumeSpecName "kube-api-access-sbv4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.852471 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn" (OuterVolumeSpecName: "kube-api-access-w7wfn") pod "d8322b3d-5eac-45bc-ad35-f98e62cf9715" (UID: "d8322b3d-5eac-45bc-ad35-f98e62cf9715"). InnerVolumeSpecName "kube-api-access-w7wfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.863967 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k" (OuterVolumeSpecName: "kube-api-access-dcn6k") pod "303dfe99-b3ef-4f14-857b-014fd13dce75" (UID: "303dfe99-b3ef-4f14-857b-014fd13dce75"). InnerVolumeSpecName "kube-api-access-dcn6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.870595 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947242 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947326 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947362 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947463 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947516 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-996k5\" (UniqueName: \"kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.947542 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc\") pod \"17d2153c-d0ae-481f-8c8c-29c602e23114\" (UID: \"17d2153c-d0ae-481f-8c8c-29c602e23114\") " Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.948522 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7wfn\" (UniqueName: \"kubernetes.io/projected/d8322b3d-5eac-45bc-ad35-f98e62cf9715-kube-api-access-w7wfn\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.948541 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcn6k\" (UniqueName: \"kubernetes.io/projected/303dfe99-b3ef-4f14-857b-014fd13dce75-kube-api-access-dcn6k\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.948551 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbv4s\" (UniqueName: \"kubernetes.io/projected/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb-kube-api-access-sbv4s\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.951587 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5" (OuterVolumeSpecName: "kube-api-access-996k5") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "kube-api-access-996k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.988375 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.989301 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.995351 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:30 crc kubenswrapper[4837]: I1001 07:25:30.999244 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.006404 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config" (OuterVolumeSpecName: "config") pod "17d2153c-d0ae-481f-8c8c-29c602e23114" (UID: "17d2153c-d0ae-481f-8c8c-29c602e23114"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050514 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-996k5\" (UniqueName: \"kubernetes.io/projected/17d2153c-d0ae-481f-8c8c-29c602e23114-kube-api-access-996k5\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050544 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050554 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050581 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050591 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.050599 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17d2153c-d0ae-481f-8c8c-29c602e23114-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.123250 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" event={"ID":"17d2153c-d0ae-481f-8c8c-29c602e23114","Type":"ContainerDied","Data":"64dead3fa2a04a252430397e51aa28994b9323f8d87efb9918f4c1700dd0e0cb"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.123488 4837 scope.go:117] "RemoveContainer" containerID="ec4455314111452568585c736a7b869b222f7ee6602cde92ca750a4103de523f" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.123587 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674cd4cb49-vrtnq" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.126513 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerStarted","Data":"bc87cee8942d823066e07a6db61e50efe64096d5d7e9e46e63586df61e4e5c2a"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.128897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerStarted","Data":"cc574cbffee90ee1ee641c1741dcc43c1f2b4b2221b367eaec963ac066577b55"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.130232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8fcf-account-create-t728v" event={"ID":"dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb","Type":"ContainerDied","Data":"1465846bd5becfd8412919b360cbb9a880ea73f93661b69a764831198c24df2a"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.130254 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1465846bd5becfd8412919b360cbb9a880ea73f93661b69a764831198c24df2a" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.130256 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8fcf-account-create-t728v" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.131653 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerStarted","Data":"f140d4a2427401013beea5445e820c81d1b8a2720cc40ac5d452f52df83146d7"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.133293 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fpv6z" event={"ID":"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c","Type":"ContainerStarted","Data":"a4bd3be67a6248edc35435d40996f9d6bc93d305b11058483bb4679159edd1be"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.133327 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fpv6z" event={"ID":"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c","Type":"ContainerStarted","Data":"0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.136881 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f4b1-account-create-lzfzs" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.136884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f4b1-account-create-lzfzs" event={"ID":"d8322b3d-5eac-45bc-ad35-f98e62cf9715","Type":"ContainerDied","Data":"39127ff4b01518f843247e0843a7458eb77af7672bfa3f2964498f9d04c76fdf"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.137049 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39127ff4b01518f843247e0843a7458eb77af7672bfa3f2964498f9d04c76fdf" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.139137 4837 generic.go:334] "Generic (PLEG): container finished" podID="201dd153-cadb-48ca-bddc-b7ba0622f39a" containerID="c3b73bdd602f1754beef682cc6ae2997909ff57156cad987858a8ff3e196d7a8" exitCode=0 Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.139182 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8fqtj" event={"ID":"201dd153-cadb-48ca-bddc-b7ba0622f39a","Type":"ContainerDied","Data":"c3b73bdd602f1754beef682cc6ae2997909ff57156cad987858a8ff3e196d7a8"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.143306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e1a9-account-create-6drcx" event={"ID":"303dfe99-b3ef-4f14-857b-014fd13dce75","Type":"ContainerDied","Data":"3309e518d42f5f5134132bcf936e06944f0c4061707e0dc96deed769cda635b8"} Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.143353 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3309e518d42f5f5134132bcf936e06944f0c4061707e0dc96deed769cda635b8" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.143354 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e1a9-account-create-6drcx" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.147654 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fpv6z" podStartSLOduration=2.147642393 podStartE2EDuration="2.147642393s" podCreationTimestamp="2025-10-01 07:25:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:31.144396132 +0000 UTC m=+1187.986003587" watchObservedRunningTime="2025-10-01 07:25:31.147642393 +0000 UTC m=+1187.989249848" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.156757 4837 scope.go:117] "RemoveContainer" containerID="0ce5c4756fd7fd10bb8c6fed6b1f61429d949908eaa333aecb37a612cb0bbc21" Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.274934 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.281430 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-674cd4cb49-vrtnq"] Oct 01 07:25:31 crc kubenswrapper[4837]: I1001 07:25:31.824962 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" path="/var/lib/kubelet/pods/17d2153c-d0ae-481f-8c8c-29c602e23114/volumes" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.162001 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerStarted","Data":"db3da1ff91657924b3e2d7ae07a63ed06b13f469b3bbe00f8607fc6501c9abd5"} Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.162055 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerStarted","Data":"a2a36e3ba660fceacc1d7e98c927ee4199206baa6f52ac1dced189d0a31882c6"} Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.171586 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerStarted","Data":"9a6a1a98f80d99ded752635627bf4357fc7c57b99d84065d9f695f84bd4c75e9"} Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.185251 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.185235116 podStartE2EDuration="3.185235116s" podCreationTimestamp="2025-10-01 07:25:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:32.183845822 +0000 UTC m=+1189.025453277" watchObservedRunningTime="2025-10-01 07:25:32.185235116 +0000 UTC m=+1189.026842571" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.223096 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.223081092 podStartE2EDuration="4.223081092s" podCreationTimestamp="2025-10-01 07:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:32.208120942 +0000 UTC m=+1189.049728397" watchObservedRunningTime="2025-10-01 07:25:32.223081092 +0000 UTC m=+1189.064688547" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336041 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-cx8fl"] Oct 01 07:25:32 crc kubenswrapper[4837]: E1001 07:25:32.336349 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="dnsmasq-dns" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336359 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="dnsmasq-dns" Oct 01 07:25:32 crc kubenswrapper[4837]: E1001 07:25:32.336371 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="init" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336377 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="init" Oct 01 07:25:32 crc kubenswrapper[4837]: E1001 07:25:32.336388 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303dfe99-b3ef-4f14-857b-014fd13dce75" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336394 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="303dfe99-b3ef-4f14-857b-014fd13dce75" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: E1001 07:25:32.336443 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8322b3d-5eac-45bc-ad35-f98e62cf9715" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336451 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8322b3d-5eac-45bc-ad35-f98e62cf9715" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: E1001 07:25:32.336464 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.336811 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.337054 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d2153c-d0ae-481f-8c8c-29c602e23114" containerName="dnsmasq-dns" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.337076 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.337087 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="303dfe99-b3ef-4f14-857b-014fd13dce75" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.337099 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8322b3d-5eac-45bc-ad35-f98e62cf9715" containerName="mariadb-account-create" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.340215 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.342816 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d5k92" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.342868 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.343105 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.346836 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cx8fl"] Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.471765 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.471823 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.471850 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52qff\" (UniqueName: \"kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.545627 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.575572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.575647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.575746 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52qff\" (UniqueName: \"kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.597507 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.599490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.603244 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52qff\" (UniqueName: \"kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff\") pod \"neutron-db-sync-cx8fl\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.670744 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678228 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs\") pod \"201dd153-cadb-48ca-bddc-b7ba0622f39a\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678288 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdtbj\" (UniqueName: \"kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj\") pod \"201dd153-cadb-48ca-bddc-b7ba0622f39a\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678314 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data\") pod \"201dd153-cadb-48ca-bddc-b7ba0622f39a\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678556 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs" (OuterVolumeSpecName: "logs") pod "201dd153-cadb-48ca-bddc-b7ba0622f39a" (UID: "201dd153-cadb-48ca-bddc-b7ba0622f39a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678664 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle\") pod \"201dd153-cadb-48ca-bddc-b7ba0622f39a\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.678821 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts\") pod \"201dd153-cadb-48ca-bddc-b7ba0622f39a\" (UID: \"201dd153-cadb-48ca-bddc-b7ba0622f39a\") " Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.679177 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/201dd153-cadb-48ca-bddc-b7ba0622f39a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.682173 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts" (OuterVolumeSpecName: "scripts") pod "201dd153-cadb-48ca-bddc-b7ba0622f39a" (UID: "201dd153-cadb-48ca-bddc-b7ba0622f39a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.682369 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj" (OuterVolumeSpecName: "kube-api-access-kdtbj") pod "201dd153-cadb-48ca-bddc-b7ba0622f39a" (UID: "201dd153-cadb-48ca-bddc-b7ba0622f39a"). InnerVolumeSpecName "kube-api-access-kdtbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.708832 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "201dd153-cadb-48ca-bddc-b7ba0622f39a" (UID: "201dd153-cadb-48ca-bddc-b7ba0622f39a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.712819 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data" (OuterVolumeSpecName: "config-data") pod "201dd153-cadb-48ca-bddc-b7ba0622f39a" (UID: "201dd153-cadb-48ca-bddc-b7ba0622f39a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.780714 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.780738 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.780747 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdtbj\" (UniqueName: \"kubernetes.io/projected/201dd153-cadb-48ca-bddc-b7ba0622f39a-kube-api-access-kdtbj\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:32 crc kubenswrapper[4837]: I1001 07:25:32.780758 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/201dd153-cadb-48ca-bddc-b7ba0622f39a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.115391 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cx8fl"] Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.200273 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8fqtj" event={"ID":"201dd153-cadb-48ca-bddc-b7ba0622f39a","Type":"ContainerDied","Data":"0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875"} Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.200420 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0955a7dc91add192daf74f952841a018f569ffe8439ff973f3c0f267d4280875" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.200504 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8fqtj" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.201881 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cx8fl" event={"ID":"7fbf0ac2-1b84-4e30-8328-31554d960caf","Type":"ContainerStarted","Data":"fd00bf2d89e9626dc89941349072369da5fdf6124863cc68d611e990fd02d860"} Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.270320 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:25:33 crc kubenswrapper[4837]: E1001 07:25:33.271017 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201dd153-cadb-48ca-bddc-b7ba0622f39a" containerName="placement-db-sync" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.271130 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="201dd153-cadb-48ca-bddc-b7ba0622f39a" containerName="placement-db-sync" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.271363 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="201dd153-cadb-48ca-bddc-b7ba0622f39a" containerName="placement-db-sync" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.272279 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.275389 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.275579 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.275705 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.275810 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.276113 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cm8kg" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.299741 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391466 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391515 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391536 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqxs9\" (UniqueName: \"kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391682 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.391724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.492923 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493031 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493062 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493131 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqxs9\" (UniqueName: \"kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493152 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.493195 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.496909 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.501387 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.510266 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.512221 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.512237 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.515673 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.520361 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqxs9\" (UniqueName: \"kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9\") pod \"placement-7cf5d6fffb-62nql\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:33 crc kubenswrapper[4837]: I1001 07:25:33.618076 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:34 crc kubenswrapper[4837]: I1001 07:25:34.221723 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cx8fl" event={"ID":"7fbf0ac2-1b84-4e30-8328-31554d960caf","Type":"ContainerStarted","Data":"1ba5135a56f71ec6e35ec8aca0609be3d9cd38f58451416cc4bbea13198a11fe"} Oct 01 07:25:34 crc kubenswrapper[4837]: I1001 07:25:34.224249 4837 generic.go:334] "Generic (PLEG): container finished" podID="9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" containerID="a4bd3be67a6248edc35435d40996f9d6bc93d305b11058483bb4679159edd1be" exitCode=0 Oct 01 07:25:34 crc kubenswrapper[4837]: I1001 07:25:34.224269 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fpv6z" event={"ID":"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c","Type":"ContainerDied","Data":"a4bd3be67a6248edc35435d40996f9d6bc93d305b11058483bb4679159edd1be"} Oct 01 07:25:34 crc kubenswrapper[4837]: I1001 07:25:34.240931 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-cx8fl" podStartSLOduration=2.240915459 podStartE2EDuration="2.240915459s" podCreationTimestamp="2025-10-01 07:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:34.239960336 +0000 UTC m=+1191.081567801" watchObservedRunningTime="2025-10-01 07:25:34.240915459 +0000 UTC m=+1191.082522914" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.839042 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.891619 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5q9zm"] Oct 01 07:25:36 crc kubenswrapper[4837]: E1001 07:25:36.891984 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" containerName="keystone-bootstrap" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.892003 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" containerName="keystone-bootstrap" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.892192 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" containerName="keystone-bootstrap" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.892743 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.896561 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.896723 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.896935 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dpp55" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.901622 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5q9zm"] Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.963923 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.963966 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5hjf\" (UniqueName: \"kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.964037 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.964094 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.964146 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.964173 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys\") pod \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\" (UID: \"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c\") " Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.972828 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.972890 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.972929 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf" (OuterVolumeSpecName: "kube-api-access-g5hjf") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "kube-api-access-g5hjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.973047 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts" (OuterVolumeSpecName: "scripts") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:36 crc kubenswrapper[4837]: I1001 07:25:36.990416 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data" (OuterVolumeSpecName: "config-data") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.005528 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" (UID: "9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.062559 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-jw42k"] Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.063842 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066158 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8472c" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066333 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066489 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066530 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066599 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzskg\" (UniqueName: \"kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066616 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066655 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066681 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066743 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066757 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066766 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066774 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066784 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5hjf\" (UniqueName: \"kubernetes.io/projected/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-kube-api-access-g5hjf\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.066793 4837 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.075338 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jw42k"] Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168383 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168434 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9js7w\" (UniqueName: \"kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168467 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168493 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168521 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168547 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168623 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzskg\" (UniqueName: \"kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.168646 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.169204 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.172484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.172521 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.172839 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.172867 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.183789 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:25:37 crc kubenswrapper[4837]: W1001 07:25:37.196420 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6212cb23_af83_4bcf_8328_91b14a62844b.slice/crio-acea766de68aecf54adb5656fbbed042c79ac7bbd52ffa173d7469fa3356ae48 WatchSource:0}: Error finding container acea766de68aecf54adb5656fbbed042c79ac7bbd52ffa173d7469fa3356ae48: Status 404 returned error can't find the container with id acea766de68aecf54adb5656fbbed042c79ac7bbd52ffa173d7469fa3356ae48 Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.199334 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzskg\" (UniqueName: \"kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg\") pod \"cinder-db-sync-5q9zm\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.227447 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.268403 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerStarted","Data":"52139a22cf4186c90e213ed726bb6ce208e70a63fb9bbeefaa0dd510a9f59d78"} Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.270153 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerStarted","Data":"acea766de68aecf54adb5656fbbed042c79ac7bbd52ffa173d7469fa3356ae48"} Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.270632 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9js7w\" (UniqueName: \"kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.270745 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.270816 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.273296 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fpv6z" event={"ID":"9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c","Type":"ContainerDied","Data":"0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa"} Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.273324 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0267d260c6f480836d41f212b4fe800ead5845326b003ed9de16b230cf8b75aa" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.273383 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fpv6z" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.274295 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.280767 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.288041 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9js7w\" (UniqueName: \"kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w\") pod \"barbican-db-sync-jw42k\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.404119 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jw42k" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.531904 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5q9zm"] Oct 01 07:25:37 crc kubenswrapper[4837]: W1001 07:25:37.559274 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80f48dbc_6a43_46a7_bca1_00027ded7d1f.slice/crio-8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb WatchSource:0}: Error finding container 8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb: Status 404 returned error can't find the container with id 8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.855532 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jw42k"] Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.962062 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.963357 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.964961 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.966723 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.967423 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.967726 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.970573 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lcpw5" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.970823 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 07:25:37 crc kubenswrapper[4837]: I1001 07:25:37.988094 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088156 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088251 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088401 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088455 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088497 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088523 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgd6l\" (UniqueName: \"kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.088557 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189707 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189771 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189798 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189837 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189864 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgd6l\" (UniqueName: \"kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189894 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189929 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.189956 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.194307 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.194870 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.196871 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.198158 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.200423 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.208825 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.209168 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.211230 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgd6l\" (UniqueName: \"kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l\") pod \"keystone-5799b746f7-mzft5\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.287185 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.300796 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q9zm" event={"ID":"80f48dbc-6a43-46a7-bca1-00027ded7d1f","Type":"ContainerStarted","Data":"8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb"} Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.305844 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jw42k" event={"ID":"d1c94750-373c-41d4-aa27-7d580f35b967","Type":"ContainerStarted","Data":"0753a8157a870f6d9b7107000963e995e721fd3cdacad46561d5dfcd26b5e5fe"} Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.308586 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerStarted","Data":"5f3d01a41efebb2a65c4e6e279d47ec4c2349d13a91fa0e7c788a05bf6c23dbd"} Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.308627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerStarted","Data":"0526bde28972f0419714df8a5e921f83607729d39715b762ffd0a2c5dd4a7841"} Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.309231 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.309277 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.329995 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7cf5d6fffb-62nql" podStartSLOduration=5.329973064 podStartE2EDuration="5.329973064s" podCreationTimestamp="2025-10-01 07:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:38.327292628 +0000 UTC m=+1195.168900083" watchObservedRunningTime="2025-10-01 07:25:38.329973064 +0000 UTC m=+1195.171580519" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.754440 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.754510 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.772487 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.826668 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:38 crc kubenswrapper[4837]: I1001 07:25:38.834574 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.319927 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5799b746f7-mzft5" event={"ID":"943a976f-044a-4b42-b08c-baeae7b85f42","Type":"ContainerStarted","Data":"700b94b97abe1787c669a876fca142041a7d8e45ca92e097e477d59f24bd52a0"} Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.320321 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.320429 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.536032 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.536081 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.595064 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 07:25:39 crc kubenswrapper[4837]: I1001 07:25:39.609592 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 07:25:40 crc kubenswrapper[4837]: I1001 07:25:40.329153 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 07:25:40 crc kubenswrapper[4837]: I1001 07:25:40.329186 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 07:25:41 crc kubenswrapper[4837]: I1001 07:25:41.141199 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:41 crc kubenswrapper[4837]: I1001 07:25:41.180770 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.354080 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5799b746f7-mzft5" event={"ID":"943a976f-044a-4b42-b08c-baeae7b85f42","Type":"ContainerStarted","Data":"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3"} Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.354352 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.381145 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5799b746f7-mzft5" podStartSLOduration=5.381116862 podStartE2EDuration="5.381116862s" podCreationTimestamp="2025-10-01 07:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:25:42.368280324 +0000 UTC m=+1199.209887779" watchObservedRunningTime="2025-10-01 07:25:42.381116862 +0000 UTC m=+1199.222724317" Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.629679 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.633074 4837 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 07:25:42 crc kubenswrapper[4837]: I1001 07:25:42.726177 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 07:25:51 crc kubenswrapper[4837]: I1001 07:25:51.439064 4837 generic.go:334] "Generic (PLEG): container finished" podID="7fbf0ac2-1b84-4e30-8328-31554d960caf" containerID="1ba5135a56f71ec6e35ec8aca0609be3d9cd38f58451416cc4bbea13198a11fe" exitCode=0 Oct 01 07:25:51 crc kubenswrapper[4837]: I1001 07:25:51.439110 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cx8fl" event={"ID":"7fbf0ac2-1b84-4e30-8328-31554d960caf","Type":"ContainerDied","Data":"1ba5135a56f71ec6e35ec8aca0609be3d9cd38f58451416cc4bbea13198a11fe"} Oct 01 07:25:56 crc kubenswrapper[4837]: E1001 07:25:56.927854 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 01 07:25:56 crc kubenswrapper[4837]: E1001 07:25:56.928513 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-82hds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9f55b0fc-aa86-4bc0-a072-03f52fa47f87): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 07:25:56 crc kubenswrapper[4837]: E1001 07:25:56.929764 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.008315 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.070712 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52qff\" (UniqueName: \"kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff\") pod \"7fbf0ac2-1b84-4e30-8328-31554d960caf\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.071108 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config\") pod \"7fbf0ac2-1b84-4e30-8328-31554d960caf\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.071307 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle\") pod \"7fbf0ac2-1b84-4e30-8328-31554d960caf\" (UID: \"7fbf0ac2-1b84-4e30-8328-31554d960caf\") " Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.080656 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff" (OuterVolumeSpecName: "kube-api-access-52qff") pod "7fbf0ac2-1b84-4e30-8328-31554d960caf" (UID: "7fbf0ac2-1b84-4e30-8328-31554d960caf"). InnerVolumeSpecName "kube-api-access-52qff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.097789 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fbf0ac2-1b84-4e30-8328-31554d960caf" (UID: "7fbf0ac2-1b84-4e30-8328-31554d960caf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.116070 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config" (OuterVolumeSpecName: "config") pod "7fbf0ac2-1b84-4e30-8328-31554d960caf" (UID: "7fbf0ac2-1b84-4e30-8328-31554d960caf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.173745 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.173780 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52qff\" (UniqueName: \"kubernetes.io/projected/7fbf0ac2-1b84-4e30-8328-31554d960caf-kube-api-access-52qff\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.173792 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbf0ac2-1b84-4e30-8328-31554d960caf-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.492943 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cx8fl" event={"ID":"7fbf0ac2-1b84-4e30-8328-31554d960caf","Type":"ContainerDied","Data":"fd00bf2d89e9626dc89941349072369da5fdf6124863cc68d611e990fd02d860"} Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.492978 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cx8fl" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.492986 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd00bf2d89e9626dc89941349072369da5fdf6124863cc68d611e990fd02d860" Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.493054 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-central-agent" containerID="cri-o://8fcfd7702a2a1f358988e7055d8a5cfae7b3a6dbbe84448d6f78c2cfbb4d95ad" gracePeriod=30 Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.493108 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="sg-core" containerID="cri-o://52139a22cf4186c90e213ed726bb6ce208e70a63fb9bbeefaa0dd510a9f59d78" gracePeriod=30 Oct 01 07:25:57 crc kubenswrapper[4837]: I1001 07:25:57.493147 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-notification-agent" containerID="cri-o://bc87cee8942d823066e07a6db61e50efe64096d5d7e9e46e63586df61e4e5c2a" gracePeriod=30 Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.038376 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695" Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.039068 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzskg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5q9zm_openstack(80f48dbc-6a43-46a7-bca1-00027ded7d1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.040264 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5q9zm" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.289170 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.290102 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbf0ac2-1b84-4e30-8328-31554d960caf" containerName="neutron-db-sync" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.290118 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbf0ac2-1b84-4e30-8328-31554d960caf" containerName="neutron-db-sync" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.298221 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbf0ac2-1b84-4e30-8328-31554d960caf" containerName="neutron-db-sync" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.300718 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.346681 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404300 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404370 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404561 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404623 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404649 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.404826 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb4qx\" (UniqueName: \"kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.432888 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.434185 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.438556 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.438593 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.438714 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d5k92" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.438798 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.479745 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8v9p\" (UniqueName: \"kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb4qx\" (UniqueName: \"kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506562 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506591 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506622 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506659 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506676 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506728 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506758 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506778 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.506794 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.507825 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.507845 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.508341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.508529 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.508936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.512349 4837 generic.go:334] "Generic (PLEG): container finished" podID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerID="52139a22cf4186c90e213ed726bb6ce208e70a63fb9bbeefaa0dd510a9f59d78" exitCode=2 Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.512380 4837 generic.go:334] "Generic (PLEG): container finished" podID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerID="8fcfd7702a2a1f358988e7055d8a5cfae7b3a6dbbe84448d6f78c2cfbb4d95ad" exitCode=0 Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.512569 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerDied","Data":"52139a22cf4186c90e213ed726bb6ce208e70a63fb9bbeefaa0dd510a9f59d78"} Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.512622 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerDied","Data":"8fcfd7702a2a1f358988e7055d8a5cfae7b3a6dbbe84448d6f78c2cfbb4d95ad"} Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.513810 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695\\\"\"" pod="openstack/cinder-db-sync-5q9zm" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.529100 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb4qx\" (UniqueName: \"kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx\") pod \"dnsmasq-dns-7c964fd46f-pbtxw\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.608162 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8v9p\" (UniqueName: \"kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.608450 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.608558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.608634 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.608732 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.613418 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.613456 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.621303 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.621471 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.635490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8v9p\" (UniqueName: \"kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p\") pod \"neutron-85544dfb8d-wtz4w\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.660092 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:25:58 crc kubenswrapper[4837]: I1001 07:25:58.754437 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.806197 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:73fd28af83ea96cc920d26dba6105ee59f0824234527949884e6ca55b71d7533" Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.806350 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:73fd28af83ea96cc920d26dba6105ee59f0824234527949884e6ca55b71d7533,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9js7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-jw42k_openstack(d1c94750-373c-41d4-aa27-7d580f35b967): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 07:25:58 crc kubenswrapper[4837]: E1001 07:25:58.808671 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-jw42k" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" Oct 01 07:25:59 crc kubenswrapper[4837]: I1001 07:25:59.508820 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:25:59 crc kubenswrapper[4837]: W1001 07:25:59.516389 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5f90d53_0db2_4d86_8bda_8ff78e80c604.slice/crio-cf731f0bb246b88e81017a974579d3ae7108ca6fb9be06051ac979cacdf90bd8 WatchSource:0}: Error finding container cf731f0bb246b88e81017a974579d3ae7108ca6fb9be06051ac979cacdf90bd8: Status 404 returned error can't find the container with id cf731f0bb246b88e81017a974579d3ae7108ca6fb9be06051ac979cacdf90bd8 Oct 01 07:25:59 crc kubenswrapper[4837]: E1001 07:25:59.519981 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:73fd28af83ea96cc920d26dba6105ee59f0824234527949884e6ca55b71d7533\\\"\"" pod="openstack/barbican-db-sync-jw42k" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" Oct 01 07:25:59 crc kubenswrapper[4837]: I1001 07:25:59.560212 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.527088 4837 generic.go:334] "Generic (PLEG): container finished" podID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerID="a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020" exitCode=0 Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.527206 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" event={"ID":"b5f90d53-0db2-4d86-8bda-8ff78e80c604","Type":"ContainerDied","Data":"a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020"} Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.527731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" event={"ID":"b5f90d53-0db2-4d86-8bda-8ff78e80c604","Type":"ContainerStarted","Data":"cf731f0bb246b88e81017a974579d3ae7108ca6fb9be06051ac979cacdf90bd8"} Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.539734 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerStarted","Data":"dd9dba269473c23311be1d264a0a8a5a066fb07a7e2e6a62e54478bad4794813"} Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.539783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerStarted","Data":"93da5dd64723a812493ca4183a35399159dcfacc340b861719f66b99aa3185a3"} Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.539797 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerStarted","Data":"a4459cbf7d3ef13f30ca7ca41949068142126760e65a776d9a5c34f25671ebd1"} Oct 01 07:26:00 crc kubenswrapper[4837]: I1001 07:26:00.539921 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.184142 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85544dfb8d-wtz4w" podStartSLOduration=3.18411169 podStartE2EDuration="3.18411169s" podCreationTimestamp="2025-10-01 07:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:00.574679577 +0000 UTC m=+1217.416287032" watchObservedRunningTime="2025-10-01 07:26:01.18411169 +0000 UTC m=+1218.025719155" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.189997 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.198188 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.208174 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.208443 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.231353 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362145 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362201 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362367 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362415 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362443 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362557 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b46vh\" (UniqueName: \"kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.362593 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.464584 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b46vh\" (UniqueName: \"kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.465807 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.465890 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.465925 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.466012 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.466068 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.466101 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.473672 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.474511 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.474909 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.475337 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.484617 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.489601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b46vh\" (UniqueName: \"kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.490908 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle\") pod \"neutron-6cf464dc69-n54w7\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.530327 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.550352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" event={"ID":"b5f90d53-0db2-4d86-8bda-8ff78e80c604","Type":"ContainerStarted","Data":"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df"} Oct 01 07:26:01 crc kubenswrapper[4837]: I1001 07:26:01.577833 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" podStartSLOduration=3.577808007 podStartE2EDuration="3.577808007s" podCreationTimestamp="2025-10-01 07:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:01.564808415 +0000 UTC m=+1218.406415880" watchObservedRunningTime="2025-10-01 07:26:01.577808007 +0000 UTC m=+1218.419415502" Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.076369 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.565373 4837 generic.go:334] "Generic (PLEG): container finished" podID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerID="bc87cee8942d823066e07a6db61e50efe64096d5d7e9e46e63586df61e4e5c2a" exitCode=0 Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.565580 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerDied","Data":"bc87cee8942d823066e07a6db61e50efe64096d5d7e9e46e63586df61e4e5c2a"} Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.569315 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerStarted","Data":"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0"} Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.569351 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.569362 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerStarted","Data":"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9"} Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.569371 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerStarted","Data":"2152ae928c2d6986b8654a591590735c5c17e4168ef13956b92d25ece300243a"} Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.569402 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.597038 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cf464dc69-n54w7" podStartSLOduration=1.597021115 podStartE2EDuration="1.597021115s" podCreationTimestamp="2025-10-01 07:26:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:02.595015826 +0000 UTC m=+1219.436623291" watchObservedRunningTime="2025-10-01 07:26:02.597021115 +0000 UTC m=+1219.438628570" Oct 01 07:26:02 crc kubenswrapper[4837]: I1001 07:26:02.906426 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.018606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82hds\" (UniqueName: \"kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019017 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019071 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019167 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019296 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.019326 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data\") pod \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\" (UID: \"9f55b0fc-aa86-4bc0-a072-03f52fa47f87\") " Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.020916 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.021135 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.026654 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts" (OuterVolumeSpecName: "scripts") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.027629 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds" (OuterVolumeSpecName: "kube-api-access-82hds") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "kube-api-access-82hds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.056816 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.075050 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data" (OuterVolumeSpecName: "config-data") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.100143 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f55b0fc-aa86-4bc0-a072-03f52fa47f87" (UID: "9f55b0fc-aa86-4bc0-a072-03f52fa47f87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121838 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121888 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121911 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82hds\" (UniqueName: \"kubernetes.io/projected/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-kube-api-access-82hds\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121934 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121953 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121970 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.121986 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f55b0fc-aa86-4bc0-a072-03f52fa47f87-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.584518 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f55b0fc-aa86-4bc0-a072-03f52fa47f87","Type":"ContainerDied","Data":"18c66ecbd2e9e0d825743ad001d1b061d1e67fd3b3598f865300922005a24f50"} Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.584605 4837 scope.go:117] "RemoveContainer" containerID="52139a22cf4186c90e213ed726bb6ce208e70a63fb9bbeefaa0dd510a9f59d78" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.584888 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.663341 4837 scope.go:117] "RemoveContainer" containerID="bc87cee8942d823066e07a6db61e50efe64096d5d7e9e46e63586df61e4e5c2a" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.706263 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.710342 4837 scope.go:117] "RemoveContainer" containerID="8fcfd7702a2a1f358988e7055d8a5cfae7b3a6dbbe84448d6f78c2cfbb4d95ad" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.744579 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.757837 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:03 crc kubenswrapper[4837]: E1001 07:26:03.759185 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-notification-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.759206 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-notification-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: E1001 07:26:03.759245 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-central-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.759252 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-central-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: E1001 07:26:03.759272 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="sg-core" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.759281 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="sg-core" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.759940 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-notification-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.759979 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="ceilometer-central-agent" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.760003 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" containerName="sg-core" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.771783 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.775496 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.776517 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.794598 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.827741 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f55b0fc-aa86-4bc0-a072-03f52fa47f87" path="/var/lib/kubelet/pods/9f55b0fc-aa86-4bc0-a072-03f52fa47f87/volumes" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.956823 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.956920 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.956964 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.957007 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55jv2\" (UniqueName: \"kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.957064 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.957237 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:03 crc kubenswrapper[4837]: I1001 07:26:03.957302 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.059960 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060497 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060560 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060633 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060732 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060775 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.060820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55jv2\" (UniqueName: \"kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.066122 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.066282 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.066375 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.069104 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.070081 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.076065 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.086446 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55jv2\" (UniqueName: \"kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2\") pod \"ceilometer-0\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.092185 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:04 crc kubenswrapper[4837]: I1001 07:26:04.586396 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:04 crc kubenswrapper[4837]: W1001 07:26:04.591237 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2887fcce_b51b_4de8_8882_2c4ad6ae5ef3.slice/crio-016fb34e2265810591a47cf2c1a62bf28edb0d93702eeff77027a1dd738d3dbc WatchSource:0}: Error finding container 016fb34e2265810591a47cf2c1a62bf28edb0d93702eeff77027a1dd738d3dbc: Status 404 returned error can't find the container with id 016fb34e2265810591a47cf2c1a62bf28edb0d93702eeff77027a1dd738d3dbc Oct 01 07:26:05 crc kubenswrapper[4837]: I1001 07:26:05.319149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:26:05 crc kubenswrapper[4837]: I1001 07:26:05.621575 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerStarted","Data":"016fb34e2265810591a47cf2c1a62bf28edb0d93702eeff77027a1dd738d3dbc"} Oct 01 07:26:06 crc kubenswrapper[4837]: I1001 07:26:06.243314 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:26:06 crc kubenswrapper[4837]: I1001 07:26:06.631177 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerStarted","Data":"8a0153df94f9f118f11c612e66a09395d7db0607109889545449400fc92dcd07"} Oct 01 07:26:06 crc kubenswrapper[4837]: I1001 07:26:06.631719 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerStarted","Data":"8569d4ef81d70b9307fd75874b5f755b4a02373d0fc0af13c4accdfa2b783889"} Oct 01 07:26:07 crc kubenswrapper[4837]: I1001 07:26:07.643865 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerStarted","Data":"b4a8c7f62d753a71e56d6720b2d6f20d8c2487eb9d0dbee04931b1295f5baf8e"} Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.652676 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerStarted","Data":"2085c52d4732f1b4e2efc0935736dd42e0576a71ae1355fb519754fc0d79c93e"} Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.654015 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.664734 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.680471 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.086380524 podStartE2EDuration="5.680456757s" podCreationTimestamp="2025-10-01 07:26:03 +0000 UTC" firstStartedPulling="2025-10-01 07:26:04.59333245 +0000 UTC m=+1221.434939905" lastFinishedPulling="2025-10-01 07:26:08.187408643 +0000 UTC m=+1225.029016138" observedRunningTime="2025-10-01 07:26:08.680173151 +0000 UTC m=+1225.521780606" watchObservedRunningTime="2025-10-01 07:26:08.680456757 +0000 UTC m=+1225.522064212" Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.769458 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:26:08 crc kubenswrapper[4837]: I1001 07:26:08.769775 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="dnsmasq-dns" containerID="cri-o://1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb" gracePeriod=10 Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.195378 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.257781 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgtzn\" (UniqueName: \"kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.257851 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.257896 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.258011 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.258028 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.258053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb\") pod \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\" (UID: \"2ee17db9-cebc-49b6-8fee-7dedb0c04622\") " Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.280289 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn" (OuterVolumeSpecName: "kube-api-access-qgtzn") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "kube-api-access-qgtzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.313893 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.360767 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.360794 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgtzn\" (UniqueName: \"kubernetes.io/projected/2ee17db9-cebc-49b6-8fee-7dedb0c04622-kube-api-access-qgtzn\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.361337 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.361414 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config" (OuterVolumeSpecName: "config") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.366198 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.366354 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ee17db9-cebc-49b6-8fee-7dedb0c04622" (UID: "2ee17db9-cebc-49b6-8fee-7dedb0c04622"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.462002 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.462035 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.462045 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.462053 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee17db9-cebc-49b6-8fee-7dedb0c04622-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.663150 4837 generic.go:334] "Generic (PLEG): container finished" podID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerID="1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb" exitCode=0 Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.663235 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.663255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" event={"ID":"2ee17db9-cebc-49b6-8fee-7dedb0c04622","Type":"ContainerDied","Data":"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb"} Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.663647 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d7cf9447-t7wtm" event={"ID":"2ee17db9-cebc-49b6-8fee-7dedb0c04622","Type":"ContainerDied","Data":"c567c0873991d3d42b8a6e3b05cbf8f2345ebb032c8dc58413999ca772c5bc58"} Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.663701 4837 scope.go:117] "RemoveContainer" containerID="1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.686114 4837 scope.go:117] "RemoveContainer" containerID="1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.695434 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.705978 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65d7cf9447-t7wtm"] Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.718808 4837 scope.go:117] "RemoveContainer" containerID="1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb" Oct 01 07:26:09 crc kubenswrapper[4837]: E1001 07:26:09.719290 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb\": container with ID starting with 1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb not found: ID does not exist" containerID="1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.719346 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb"} err="failed to get container status \"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb\": rpc error: code = NotFound desc = could not find container \"1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb\": container with ID starting with 1a0db8e8d3430ca032f1b62e2160ee51fc1f9612e925921a23ac6739a93fa4eb not found: ID does not exist" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.719377 4837 scope.go:117] "RemoveContainer" containerID="1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0" Oct 01 07:26:09 crc kubenswrapper[4837]: E1001 07:26:09.719744 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0\": container with ID starting with 1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0 not found: ID does not exist" containerID="1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.719792 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0"} err="failed to get container status \"1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0\": rpc error: code = NotFound desc = could not find container \"1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0\": container with ID starting with 1b4a270ca401805f08dd98b8740e0511fbeb362f3edc48e82a7591ad081dfcb0 not found: ID does not exist" Oct 01 07:26:09 crc kubenswrapper[4837]: I1001 07:26:09.827779 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" path="/var/lib/kubelet/pods/2ee17db9-cebc-49b6-8fee-7dedb0c04622/volumes" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.030070 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.931096 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 07:26:10 crc kubenswrapper[4837]: E1001 07:26:10.931951 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="init" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.931968 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="init" Oct 01 07:26:10 crc kubenswrapper[4837]: E1001 07:26:10.932005 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="dnsmasq-dns" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.932013 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="dnsmasq-dns" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.932241 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee17db9-cebc-49b6-8fee-7dedb0c04622" containerName="dnsmasq-dns" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.932972 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.938701 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-29ncg" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.939755 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.940777 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.947286 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.992575 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.992639 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.992751 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:10 crc kubenswrapper[4837]: I1001 07:26:10.992772 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w67zv\" (UniqueName: \"kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.094408 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.094539 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.094567 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w67zv\" (UniqueName: \"kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.094635 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.095967 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.100885 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.103163 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.119926 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w67zv\" (UniqueName: \"kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv\") pod \"openstackclient\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.256212 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.704099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jw42k" event={"ID":"d1c94750-373c-41d4-aa27-7d580f35b967","Type":"ContainerStarted","Data":"f2f44d90065f3b09e2e0423a1cb4fbaf07ef26a225a477036e37641545512570"} Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.746512 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-jw42k" podStartSLOduration=1.254913248 podStartE2EDuration="34.74649487s" podCreationTimestamp="2025-10-01 07:25:37 +0000 UTC" firstStartedPulling="2025-10-01 07:25:37.862070222 +0000 UTC m=+1194.703677687" lastFinishedPulling="2025-10-01 07:26:11.353651864 +0000 UTC m=+1228.195259309" observedRunningTime="2025-10-01 07:26:11.72022258 +0000 UTC m=+1228.561830055" watchObservedRunningTime="2025-10-01 07:26:11.74649487 +0000 UTC m=+1228.588102315" Oct 01 07:26:11 crc kubenswrapper[4837]: I1001 07:26:11.750917 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 07:26:12 crc kubenswrapper[4837]: I1001 07:26:12.715886 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc","Type":"ContainerStarted","Data":"0efc871afe7ef5ab8dff3d55c43ae710e2a28485ece07db157025caf8fe5fdf9"} Oct 01 07:26:13 crc kubenswrapper[4837]: I1001 07:26:13.727959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q9zm" event={"ID":"80f48dbc-6a43-46a7-bca1-00027ded7d1f","Type":"ContainerStarted","Data":"f94c2451dd653c31792c302868b2003480f76be40f9c075f3e63a6788149c348"} Oct 01 07:26:13 crc kubenswrapper[4837]: I1001 07:26:13.744517 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5q9zm" podStartSLOduration=3.031995838 podStartE2EDuration="37.744503697s" podCreationTimestamp="2025-10-01 07:25:36 +0000 UTC" firstStartedPulling="2025-10-01 07:25:37.565606609 +0000 UTC m=+1194.407214054" lastFinishedPulling="2025-10-01 07:26:12.278114458 +0000 UTC m=+1229.119721913" observedRunningTime="2025-10-01 07:26:13.741741159 +0000 UTC m=+1230.583348614" watchObservedRunningTime="2025-10-01 07:26:13.744503697 +0000 UTC m=+1230.586111152" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.068448 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.070020 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.072215 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.073600 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.073759 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.079464 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8q4z\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141387 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141410 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141513 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141581 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141754 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.141809 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.247771 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8q4z\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.248557 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252033 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252134 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252176 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252288 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252409 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252537 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252556 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.252988 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.257009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.257876 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.260287 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.262433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.262568 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8q4z\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.272618 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs\") pod \"swift-proxy-869997d995-swqv9\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.391220 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.737605 4837 generic.go:334] "Generic (PLEG): container finished" podID="d1c94750-373c-41d4-aa27-7d580f35b967" containerID="f2f44d90065f3b09e2e0423a1cb4fbaf07ef26a225a477036e37641545512570" exitCode=0 Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.737648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jw42k" event={"ID":"d1c94750-373c-41d4-aa27-7d580f35b967","Type":"ContainerDied","Data":"f2f44d90065f3b09e2e0423a1cb4fbaf07ef26a225a477036e37641545512570"} Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.907683 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.908094 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-central-agent" containerID="cri-o://8a0153df94f9f118f11c612e66a09395d7db0607109889545449400fc92dcd07" gracePeriod=30 Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.908193 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="sg-core" containerID="cri-o://b4a8c7f62d753a71e56d6720b2d6f20d8c2487eb9d0dbee04931b1295f5baf8e" gracePeriod=30 Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.908224 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-notification-agent" containerID="cri-o://8569d4ef81d70b9307fd75874b5f755b4a02373d0fc0af13c4accdfa2b783889" gracePeriod=30 Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.908181 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="proxy-httpd" containerID="cri-o://2085c52d4732f1b4e2efc0935736dd42e0576a71ae1355fb519754fc0d79c93e" gracePeriod=30 Oct 01 07:26:14 crc kubenswrapper[4837]: I1001 07:26:14.953437 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:26:14 crc kubenswrapper[4837]: W1001 07:26:14.957705 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d39fc78_2eb7_4a57_8bbc_11733cbb366a.slice/crio-d76ca20d8e415d77e2049d146b84bc4e6beda0554fdc6547178b387ec6891ccd WatchSource:0}: Error finding container d76ca20d8e415d77e2049d146b84bc4e6beda0554fdc6547178b387ec6891ccd: Status 404 returned error can't find the container with id d76ca20d8e415d77e2049d146b84bc4e6beda0554fdc6547178b387ec6891ccd Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.748903 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerStarted","Data":"5d1ccc4bac17d9e48187ea2dd17c1341921ed55167a7aeb5bfba6f2afe624a09"} Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.749326 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.749338 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerStarted","Data":"c8bdd89e1b73c7441649e0ed8d304c53f3b5ed251d4c5fd8292eb1d03c32ff83"} Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.749348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerStarted","Data":"d76ca20d8e415d77e2049d146b84bc4e6beda0554fdc6547178b387ec6891ccd"} Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753007 4837 generic.go:334] "Generic (PLEG): container finished" podID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerID="2085c52d4732f1b4e2efc0935736dd42e0576a71ae1355fb519754fc0d79c93e" exitCode=0 Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753037 4837 generic.go:334] "Generic (PLEG): container finished" podID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerID="b4a8c7f62d753a71e56d6720b2d6f20d8c2487eb9d0dbee04931b1295f5baf8e" exitCode=2 Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753047 4837 generic.go:334] "Generic (PLEG): container finished" podID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerID="8a0153df94f9f118f11c612e66a09395d7db0607109889545449400fc92dcd07" exitCode=0 Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753078 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerDied","Data":"2085c52d4732f1b4e2efc0935736dd42e0576a71ae1355fb519754fc0d79c93e"} Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753116 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerDied","Data":"b4a8c7f62d753a71e56d6720b2d6f20d8c2487eb9d0dbee04931b1295f5baf8e"} Oct 01 07:26:15 crc kubenswrapper[4837]: I1001 07:26:15.753126 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerDied","Data":"8a0153df94f9f118f11c612e66a09395d7db0607109889545449400fc92dcd07"} Oct 01 07:26:16 crc kubenswrapper[4837]: I1001 07:26:16.764000 4837 generic.go:334] "Generic (PLEG): container finished" podID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerID="8569d4ef81d70b9307fd75874b5f755b4a02373d0fc0af13c4accdfa2b783889" exitCode=0 Oct 01 07:26:16 crc kubenswrapper[4837]: I1001 07:26:16.764799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerDied","Data":"8569d4ef81d70b9307fd75874b5f755b4a02373d0fc0af13c4accdfa2b783889"} Oct 01 07:26:16 crc kubenswrapper[4837]: I1001 07:26:16.764863 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:17 crc kubenswrapper[4837]: I1001 07:26:17.774628 4837 generic.go:334] "Generic (PLEG): container finished" podID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" containerID="f94c2451dd653c31792c302868b2003480f76be40f9c075f3e63a6788149c348" exitCode=0 Oct 01 07:26:17 crc kubenswrapper[4837]: I1001 07:26:17.774729 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q9zm" event={"ID":"80f48dbc-6a43-46a7-bca1-00027ded7d1f","Type":"ContainerDied","Data":"f94c2451dd653c31792c302868b2003480f76be40f9c075f3e63a6788149c348"} Oct 01 07:26:17 crc kubenswrapper[4837]: I1001 07:26:17.790685 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-869997d995-swqv9" podStartSLOduration=3.790669982 podStartE2EDuration="3.790669982s" podCreationTimestamp="2025-10-01 07:26:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:15.774155766 +0000 UTC m=+1232.615763221" watchObservedRunningTime="2025-10-01 07:26:17.790669982 +0000 UTC m=+1234.632277437" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.269060 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jw42k" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.274322 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.285227 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle\") pod \"d1c94750-373c-41d4-aa27-7d580f35b967\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.285341 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9js7w\" (UniqueName: \"kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w\") pod \"d1c94750-373c-41d4-aa27-7d580f35b967\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.285531 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data\") pod \"d1c94750-373c-41d4-aa27-7d580f35b967\" (UID: \"d1c94750-373c-41d4-aa27-7d580f35b967\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.294920 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w" (OuterVolumeSpecName: "kube-api-access-9js7w") pod "d1c94750-373c-41d4-aa27-7d580f35b967" (UID: "d1c94750-373c-41d4-aa27-7d580f35b967"). InnerVolumeSpecName "kube-api-access-9js7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.306324 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.311891 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d1c94750-373c-41d4-aa27-7d580f35b967" (UID: "d1c94750-373c-41d4-aa27-7d580f35b967"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.338258 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1c94750-373c-41d4-aa27-7d580f35b967" (UID: "d1c94750-373c-41d4-aa27-7d580f35b967"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387483 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387544 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387565 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55jv2\" (UniqueName: \"kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387602 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387630 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387700 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387748 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387770 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387828 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387853 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzskg\" (UniqueName: \"kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387871 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts\") pod \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\" (UID: \"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387892 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data\") pod \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\" (UID: \"80f48dbc-6a43-46a7-bca1-00027ded7d1f\") " Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.387883 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.388170 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80f48dbc-6a43-46a7-bca1-00027ded7d1f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.388181 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.388189 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1c94750-373c-41d4-aa27-7d580f35b967-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.388197 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9js7w\" (UniqueName: \"kubernetes.io/projected/d1c94750-373c-41d4-aa27-7d580f35b967-kube-api-access-9js7w\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.389418 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.389438 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.390898 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2" (OuterVolumeSpecName: "kube-api-access-55jv2") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "kube-api-access-55jv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.391532 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.392081 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts" (OuterVolumeSpecName: "scripts") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.393322 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg" (OuterVolumeSpecName: "kube-api-access-bzskg") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "kube-api-access-bzskg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.393784 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts" (OuterVolumeSpecName: "scripts") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.411180 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.411818 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.445811 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data" (OuterVolumeSpecName: "config-data") pod "80f48dbc-6a43-46a7-bca1-00027ded7d1f" (UID: "80f48dbc-6a43-46a7-bca1-00027ded7d1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.459513 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.489999 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490034 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490049 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490061 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490073 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzskg\" (UniqueName: \"kubernetes.io/projected/80f48dbc-6a43-46a7-bca1-00027ded7d1f-kube-api-access-bzskg\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490085 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490096 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490108 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490118 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490132 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55jv2\" (UniqueName: \"kubernetes.io/projected/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-kube-api-access-55jv2\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.490144 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/80f48dbc-6a43-46a7-bca1-00027ded7d1f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.498312 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data" (OuterVolumeSpecName: "config-data") pod "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" (UID: "2887fcce-b51b-4de8-8882-2c4ad6ae5ef3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.591971 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.814454 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.814513 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2887fcce-b51b-4de8-8882-2c4ad6ae5ef3","Type":"ContainerDied","Data":"016fb34e2265810591a47cf2c1a62bf28edb0d93702eeff77027a1dd738d3dbc"} Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.814652 4837 scope.go:117] "RemoveContainer" containerID="2085c52d4732f1b4e2efc0935736dd42e0576a71ae1355fb519754fc0d79c93e" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.821916 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jw42k" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.825069 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5q9zm" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.834939 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc","Type":"ContainerStarted","Data":"c5da1cb2cb85875baa579d1f9fd8e420c616437d114859950730be4875ac535e"} Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.836797 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jw42k" event={"ID":"d1c94750-373c-41d4-aa27-7d580f35b967","Type":"ContainerDied","Data":"0753a8157a870f6d9b7107000963e995e721fd3cdacad46561d5dfcd26b5e5fe"} Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.836838 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0753a8157a870f6d9b7107000963e995e721fd3cdacad46561d5dfcd26b5e5fe" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.836857 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5q9zm" event={"ID":"80f48dbc-6a43-46a7-bca1-00027ded7d1f","Type":"ContainerDied","Data":"8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb"} Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.836875 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8342c230b499aa951cd11f60c50acafef756c1149d99c5fc2c3898da7e02e2eb" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.846803 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.463683648 podStartE2EDuration="11.846786972s" podCreationTimestamp="2025-10-01 07:26:10 +0000 UTC" firstStartedPulling="2025-10-01 07:26:11.758017945 +0000 UTC m=+1228.599625410" lastFinishedPulling="2025-10-01 07:26:21.141121249 +0000 UTC m=+1237.982728734" observedRunningTime="2025-10-01 07:26:21.845279004 +0000 UTC m=+1238.686886499" watchObservedRunningTime="2025-10-01 07:26:21.846786972 +0000 UTC m=+1238.688394427" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.863978 4837 scope.go:117] "RemoveContainer" containerID="b4a8c7f62d753a71e56d6720b2d6f20d8c2487eb9d0dbee04931b1295f5baf8e" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.893125 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.896031 4837 scope.go:117] "RemoveContainer" containerID="8569d4ef81d70b9307fd75874b5f755b4a02373d0fc0af13c4accdfa2b783889" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.899866 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.919602 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.919969 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="sg-core" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.919988 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="sg-core" Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.920005 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" containerName="barbican-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920011 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" containerName="barbican-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.920021 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" containerName="cinder-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920027 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" containerName="cinder-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.920038 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="proxy-httpd" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920044 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="proxy-httpd" Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.920062 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-central-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920068 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-central-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: E1001 07:26:21.920083 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-notification-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920089 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-notification-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920248 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" containerName="barbican-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920262 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" containerName="cinder-db-sync" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920270 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-central-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920281 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="ceilometer-notification-agent" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920293 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="proxy-httpd" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.920304 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" containerName="sg-core" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.921722 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.926954 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.927124 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.938505 4837 scope.go:117] "RemoveContainer" containerID="8a0153df94f9f118f11c612e66a09395d7db0607109889545449400fc92dcd07" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.953875 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.999432 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.999468 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:21 crc kubenswrapper[4837]: I1001 07:26:21.999491 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:21.999513 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:21.999547 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:21.999570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:21.999630 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcpk4\" (UniqueName: \"kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101299 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101357 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101393 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101471 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101504 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.101588 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcpk4\" (UniqueName: \"kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.102781 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.103030 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.106047 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.106682 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.107346 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.107616 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.122964 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcpk4\" (UniqueName: \"kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4\") pod \"ceilometer-0\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.245184 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.544101 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.545651 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.548553 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.549793 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.554790 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.554936 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.555036 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8472c" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.555170 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.561853 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.574078 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.575260 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.577384 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.584090 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dpp55" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.584844 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.587479 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.621498 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.693121 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727649 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42k4j\" (UniqueName: \"kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727699 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727717 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727749 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727778 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727817 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727838 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727854 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727872 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmckq\" (UniqueName: \"kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727901 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727920 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727973 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.727992 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.728019 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9fhc\" (UniqueName: \"kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.792275 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.796091 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.813895 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.828988 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829051 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829073 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829090 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmckq\" (UniqueName: \"kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829105 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829125 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829147 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829169 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829192 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829218 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9fhc\" (UniqueName: \"kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829253 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42k4j\" (UniqueName: \"kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829269 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829285 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.829337 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.831383 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.831485 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.832236 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.832559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.837589 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.839871 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.840119 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.842831 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.843548 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.844444 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.848002 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.848190 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerStarted","Data":"9bf625bc68be2dd86f9835ab660deb00dc1235d8083b69d5469eec1c7f584eb5"} Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.849834 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.850799 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.852519 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmckq\" (UniqueName: \"kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.852853 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42k4j\" (UniqueName: \"kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j\") pod \"cinder-scheduler-0\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.855054 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data\") pod \"barbican-worker-68b46954cf-msjzx\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.857198 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.860274 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.862149 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9fhc\" (UniqueName: \"kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc\") pod \"barbican-keystone-listener-849c79f7db-h5jwx\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.863753 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.884030 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.886678 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.920465 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.922119 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.932907 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.933845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.933890 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z8n7\" (UniqueName: \"kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.933931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.933981 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.934004 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.934055 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.934678 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.955950 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:22 crc kubenswrapper[4837]: I1001 07:26:22.964034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035337 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035367 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035406 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035457 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035491 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035511 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035540 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035558 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035578 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z8n7\" (UniqueName: \"kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035634 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035656 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.035678 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.036332 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.036354 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd2zg\" (UniqueName: \"kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.036374 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wglnl\" (UniqueName: \"kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.037354 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.037875 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.038345 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.039165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.040451 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.067401 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z8n7\" (UniqueName: \"kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7\") pod \"dnsmasq-dns-6797bbc6b9-2p4ll\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.139788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.141221 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.141305 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd2zg\" (UniqueName: \"kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.141500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wglnl\" (UniqueName: \"kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.141592 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.141718 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.142611 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143448 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143709 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143805 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143944 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.143080 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.142058 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.158278 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.160173 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.160554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.175386 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.175902 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.176363 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.176359 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.177225 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.188407 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wglnl\" (UniqueName: \"kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl\") pod \"barbican-api-7b5db87b6d-sq97p\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.188433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd2zg\" (UniqueName: \"kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg\") pod \"cinder-api-0\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.190534 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.247833 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.643594 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:26:23 crc kubenswrapper[4837]: W1001 07:26:23.652062 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f871dbb_1e64_411b_acfa_39ed6b5b66ed.slice/crio-ab36d941091a7e924c9f3355720ded0ef841331ab576500130be7e7286ab52c5 WatchSource:0}: Error finding container ab36d941091a7e924c9f3355720ded0ef841331ab576500130be7e7286ab52c5: Status 404 returned error can't find the container with id ab36d941091a7e924c9f3355720ded0ef841331ab576500130be7e7286ab52c5 Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.656666 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.882923 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2887fcce-b51b-4de8-8882-2c4ad6ae5ef3" path="/var/lib/kubelet/pods/2887fcce-b51b-4de8-8882-2c4ad6ae5ef3/volumes" Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.884394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerStarted","Data":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.884440 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.884459 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.884469 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.884736 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerStarted","Data":"261b354f4530c354587c143c410bad9980c3e53191e0942f04fc35787242aa02"} Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.887200 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerStarted","Data":"dd92804295a499b6034fc83afb8968bbde241ba43cbace5d43dac9b0edf210b2"} Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.888040 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerStarted","Data":"ab36d941091a7e924c9f3355720ded0ef841331ab576500130be7e7286ab52c5"} Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.888849 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerStarted","Data":"eab354248b04ca89f90412b44f455ea4155cdc7962efcd3b71b4bfdfe58cbb67"} Oct 01 07:26:23 crc kubenswrapper[4837]: I1001 07:26:23.946976 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.398221 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.402472 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.900852 4837 generic.go:334] "Generic (PLEG): container finished" podID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerID="56bc8d2d9f6ccd18df163e5e792c85f317dd163e845422f27facf9060502b2b5" exitCode=0 Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.900898 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" event={"ID":"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2","Type":"ContainerDied","Data":"56bc8d2d9f6ccd18df163e5e792c85f317dd163e845422f27facf9060502b2b5"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.901316 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" event={"ID":"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2","Type":"ContainerStarted","Data":"8bf08a7d2853d77f5f609a6b3474f33a557bc5e3ed700d5f9587c332e4276bc7"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.914438 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerStarted","Data":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.919341 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerStarted","Data":"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.922212 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.922324 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerStarted","Data":"a43a2453a9aa618de3b3b1731ec80890fa3fce06b07ec1ade8dec57c811ed3af"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.922343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerStarted","Data":"e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.922355 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerStarted","Data":"12ec459428014c7466214d2c99d3446b10ae3bcac80009aa60ef4f936e7fd003"} Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.922372 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:24 crc kubenswrapper[4837]: I1001 07:26:24.948531 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b5db87b6d-sq97p" podStartSLOduration=2.948515927 podStartE2EDuration="2.948515927s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:24.941899313 +0000 UTC m=+1241.783506768" watchObservedRunningTime="2025-10-01 07:26:24.948515927 +0000 UTC m=+1241.790123382" Oct 01 07:26:25 crc kubenswrapper[4837]: I1001 07:26:25.943120 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" event={"ID":"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2","Type":"ContainerStarted","Data":"c5346b4bc62a9eb24a611ae5cf881437a77d4a6be9f5e881ba9a93a20c875511"} Oct 01 07:26:25 crc kubenswrapper[4837]: I1001 07:26:25.978852 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" podStartSLOduration=3.97881631 podStartE2EDuration="3.97881631s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:25.963650675 +0000 UTC m=+1242.805258130" watchObservedRunningTime="2025-10-01 07:26:25.97881631 +0000 UTC m=+1242.820423765" Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.322649 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.971787 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerStarted","Data":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.976412 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerStarted","Data":"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402"} Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.976528 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.978077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerStarted","Data":"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666"} Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.988155 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerStarted","Data":"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144"} Oct 01 07:26:26 crc kubenswrapper[4837]: I1001 07:26:26.988295 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:27 crc kubenswrapper[4837]: I1001 07:26:27.001168 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.001143495 podStartE2EDuration="5.001143495s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:26.997141436 +0000 UTC m=+1243.838748901" watchObservedRunningTime="2025-10-01 07:26:27.001143495 +0000 UTC m=+1243.842750960" Oct 01 07:26:27 crc kubenswrapper[4837]: I1001 07:26:27.260272 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.001344 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerStarted","Data":"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41"} Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.004575 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerStarted","Data":"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290"} Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.004631 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerStarted","Data":"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc"} Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.010800 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerStarted","Data":"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164"} Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.010934 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api-log" containerID="cri-o://f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" gracePeriod=30 Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.011170 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api" containerID="cri-o://c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" gracePeriod=30 Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.020444 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68b46954cf-msjzx" podStartSLOduration=3.170502298 podStartE2EDuration="6.020431766s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="2025-10-01 07:26:23.650932164 +0000 UTC m=+1240.492539619" lastFinishedPulling="2025-10-01 07:26:26.500861632 +0000 UTC m=+1243.342469087" observedRunningTime="2025-10-01 07:26:28.016048157 +0000 UTC m=+1244.857655612" watchObservedRunningTime="2025-10-01 07:26:28.020431766 +0000 UTC m=+1244.862039221" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.037897 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.386552551 podStartE2EDuration="6.037878837s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="2025-10-01 07:26:23.849840033 +0000 UTC m=+1240.691447488" lastFinishedPulling="2025-10-01 07:26:26.501166319 +0000 UTC m=+1243.342773774" observedRunningTime="2025-10-01 07:26:28.037424826 +0000 UTC m=+1244.879032281" watchObservedRunningTime="2025-10-01 07:26:28.037878837 +0000 UTC m=+1244.879486292" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.054357 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" podStartSLOduration=3.214138227 podStartE2EDuration="6.054341744s" podCreationTimestamp="2025-10-01 07:26:22 +0000 UTC" firstStartedPulling="2025-10-01 07:26:23.66330602 +0000 UTC m=+1240.504913475" lastFinishedPulling="2025-10-01 07:26:26.503509527 +0000 UTC m=+1243.345116992" observedRunningTime="2025-10-01 07:26:28.052513039 +0000 UTC m=+1244.894120494" watchObservedRunningTime="2025-10-01 07:26:28.054341744 +0000 UTC m=+1244.895949199" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.647042 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.648831 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.651072 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.651231 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.655702 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695642 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695760 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695801 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695855 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695900 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2brw\" (UniqueName: \"kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.695929 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.765075 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797079 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797138 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2brw\" (UniqueName: \"kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797173 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797212 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797237 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797266 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797314 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.797967 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.802954 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.804229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.804589 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.806377 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.817433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.851364 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2brw\" (UniqueName: \"kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw\") pod \"barbican-api-66dc5bc7b8-ttchd\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.905335 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:28 crc kubenswrapper[4837]: I1001 07:26:28.968061 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002183 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002239 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002305 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002337 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002360 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd2zg\" (UniqueName: \"kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002397 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.002427 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle\") pod \"17aed973-6f42-4564-b8fc-61b5799e7f99\" (UID: \"17aed973-6f42-4564-b8fc-61b5799e7f99\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.008618 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.008829 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs" (OuterVolumeSpecName: "logs") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.008858 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.015124 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg" (OuterVolumeSpecName: "kube-api-access-jd2zg") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "kube-api-access-jd2zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.020398 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts" (OuterVolumeSpecName: "scripts") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.033996 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerStarted","Data":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.034230 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-central-agent" containerID="cri-o://ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" gracePeriod=30 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.034328 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.034631 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="proxy-httpd" containerID="cri-o://d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" gracePeriod=30 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.034670 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="sg-core" containerID="cri-o://367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" gracePeriod=30 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.034720 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-notification-agent" containerID="cri-o://f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" gracePeriod=30 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.038707 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057230 4837 generic.go:334] "Generic (PLEG): container finished" podID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerID="c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" exitCode=0 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057259 4837 generic.go:334] "Generic (PLEG): container finished" podID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerID="f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" exitCode=143 Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057304 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerDied","Data":"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402"} Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerDied","Data":"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13"} Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057360 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"17aed973-6f42-4564-b8fc-61b5799e7f99","Type":"ContainerDied","Data":"a43a2453a9aa618de3b3b1731ec80890fa3fce06b07ec1ade8dec57c811ed3af"} Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057376 4837 scope.go:117] "RemoveContainer" containerID="c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.057552 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.077589 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.966608362 podStartE2EDuration="8.077572793s" podCreationTimestamp="2025-10-01 07:26:21 +0000 UTC" firstStartedPulling="2025-10-01 07:26:22.789266172 +0000 UTC m=+1239.630873627" lastFinishedPulling="2025-10-01 07:26:27.900230603 +0000 UTC m=+1244.741838058" observedRunningTime="2025-10-01 07:26:29.071598255 +0000 UTC m=+1245.913205710" watchObservedRunningTime="2025-10-01 07:26:29.077572793 +0000 UTC m=+1245.919180248" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.097193 4837 scope.go:117] "RemoveContainer" containerID="f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107435 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17aed973-6f42-4564-b8fc-61b5799e7f99-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107467 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17aed973-6f42-4564-b8fc-61b5799e7f99-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107478 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107486 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd2zg\" (UniqueName: \"kubernetes.io/projected/17aed973-6f42-4564-b8fc-61b5799e7f99-kube-api-access-jd2zg\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107495 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.107504 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.125959 4837 scope.go:117] "RemoveContainer" containerID="c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" Oct 01 07:26:29 crc kubenswrapper[4837]: E1001 07:26:29.126883 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402\": container with ID starting with c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402 not found: ID does not exist" containerID="c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.126941 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402"} err="failed to get container status \"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402\": rpc error: code = NotFound desc = could not find container \"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402\": container with ID starting with c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402 not found: ID does not exist" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.126967 4837 scope.go:117] "RemoveContainer" containerID="f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" Oct 01 07:26:29 crc kubenswrapper[4837]: E1001 07:26:29.127633 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13\": container with ID starting with f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13 not found: ID does not exist" containerID="f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.127651 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13"} err="failed to get container status \"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13\": rpc error: code = NotFound desc = could not find container \"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13\": container with ID starting with f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13 not found: ID does not exist" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.127666 4837 scope.go:117] "RemoveContainer" containerID="c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.128807 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402"} err="failed to get container status \"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402\": rpc error: code = NotFound desc = could not find container \"c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402\": container with ID starting with c8af4287b7117d8b1cca900e813978645205a44d24ec7e061d059e65de653402 not found: ID does not exist" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.128847 4837 scope.go:117] "RemoveContainer" containerID="f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.129136 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13"} err="failed to get container status \"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13\": rpc error: code = NotFound desc = could not find container \"f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13\": container with ID starting with f4b52cec56b0fa40e72e474af3ea9f0a0f57340604d90c5a456550dc48964f13 not found: ID does not exist" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.131798 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data" (OuterVolumeSpecName: "config-data") pod "17aed973-6f42-4564-b8fc-61b5799e7f99" (UID: "17aed973-6f42-4564-b8fc-61b5799e7f99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.209803 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17aed973-6f42-4564-b8fc-61b5799e7f99-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.389305 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.408018 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.439761 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:29 crc kubenswrapper[4837]: E1001 07:26:29.440202 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api-log" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.440222 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api-log" Oct 01 07:26:29 crc kubenswrapper[4837]: E1001 07:26:29.440240 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.440246 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.440405 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.440430 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" containerName="cinder-api-log" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.441422 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.442901 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.444547 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.449802 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.457366 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.517807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518117 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518144 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9t4g\" (UniqueName: \"kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518198 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518307 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518374 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.518406 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.545055 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.621662 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.621827 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.621949 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622033 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622128 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622244 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622336 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622421 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9t4g\" (UniqueName: \"kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.622508 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.629649 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.629766 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.638262 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.638390 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.638383 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.639873 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.643833 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.644343 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.648133 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9t4g\" (UniqueName: \"kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g\") pod \"cinder-api-0\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.791095 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.830887 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17aed973-6f42-4564-b8fc-61b5799e7f99" path="/var/lib/kubelet/pods/17aed973-6f42-4564-b8fc-61b5799e7f99/volumes" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.835044 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930075 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930502 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcpk4\" (UniqueName: \"kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930542 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930565 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930614 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930695 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930748 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.930841 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd\") pod \"e96695ac-c874-45ef-b080-45a0adce834f\" (UID: \"e96695ac-c874-45ef-b080-45a0adce834f\") " Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.931274 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.931496 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.935242 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts" (OuterVolumeSpecName: "scripts") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.939519 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4" (OuterVolumeSpecName: "kube-api-access-kcpk4") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "kube-api-access-kcpk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:29 crc kubenswrapper[4837]: I1001 07:26:29.965899 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.017684 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.034982 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.035014 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.035024 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.035033 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e96695ac-c874-45ef-b080-45a0adce834f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.035043 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcpk4\" (UniqueName: \"kubernetes.io/projected/e96695ac-c874-45ef-b080-45a0adce834f-kube-api-access-kcpk4\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068368 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerStarted","Data":"9d0c63a56b505c323247613cb05d39bfe8227df3c32c1ff6ebcabfa1e90233e6"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerStarted","Data":"b7792f2f321c19d1eb1568ae96eddc87b77661b3d811ceac5cfe3f464d5b9726"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068423 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerStarted","Data":"28ca7e3ea0beaee25f8c27512f8e60012309156a94215dabf219cd2daca357c5"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068779 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068804 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.068794 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data" (OuterVolumeSpecName: "config-data") pod "e96695ac-c874-45ef-b080-45a0adce834f" (UID: "e96695ac-c874-45ef-b080-45a0adce834f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.078952 4837 generic.go:334] "Generic (PLEG): container finished" podID="e96695ac-c874-45ef-b080-45a0adce834f" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" exitCode=0 Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.078981 4837 generic.go:334] "Generic (PLEG): container finished" podID="e96695ac-c874-45ef-b080-45a0adce834f" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" exitCode=2 Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.078990 4837 generic.go:334] "Generic (PLEG): container finished" podID="e96695ac-c874-45ef-b080-45a0adce834f" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" exitCode=0 Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.078999 4837 generic.go:334] "Generic (PLEG): container finished" podID="e96695ac-c874-45ef-b080-45a0adce834f" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" exitCode=0 Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079031 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079069 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerDied","Data":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079094 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerDied","Data":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079103 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerDied","Data":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079114 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerDied","Data":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079122 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e96695ac-c874-45ef-b080-45a0adce834f","Type":"ContainerDied","Data":"9bf625bc68be2dd86f9835ab660deb00dc1235d8083b69d5469eec1c7f584eb5"} Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.079138 4837 scope.go:117] "RemoveContainer" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.096484 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-66dc5bc7b8-ttchd" podStartSLOduration=2.096469054 podStartE2EDuration="2.096469054s" podCreationTimestamp="2025-10-01 07:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:30.086117897 +0000 UTC m=+1246.927725352" watchObservedRunningTime="2025-10-01 07:26:30.096469054 +0000 UTC m=+1246.938076509" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.107297 4837 scope.go:117] "RemoveContainer" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.131556 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.142819 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.146701 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e96695ac-c874-45ef-b080-45a0adce834f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.153876 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.154302 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="proxy-httpd" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154321 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="proxy-httpd" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.154333 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="sg-core" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154340 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="sg-core" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.154368 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-central-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154376 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-central-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.154388 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-notification-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154393 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-notification-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154560 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="sg-core" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154581 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-notification-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154594 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="ceilometer-central-agent" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.154608 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96695ac-c874-45ef-b080-45a0adce834f" containerName="proxy-httpd" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.156549 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.156665 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.161039 4837 scope.go:117] "RemoveContainer" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.161320 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.161393 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.197261 4837 scope.go:117] "RemoveContainer" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.227354 4837 scope.go:117] "RemoveContainer" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.227861 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": container with ID starting with d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e not found: ID does not exist" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.227901 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} err="failed to get container status \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": rpc error: code = NotFound desc = could not find container \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": container with ID starting with d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.227920 4837 scope.go:117] "RemoveContainer" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.228292 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": container with ID starting with 367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58 not found: ID does not exist" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.228418 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} err="failed to get container status \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": rpc error: code = NotFound desc = could not find container \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": container with ID starting with 367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.228449 4837 scope.go:117] "RemoveContainer" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.229049 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": container with ID starting with f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d not found: ID does not exist" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229102 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} err="failed to get container status \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": rpc error: code = NotFound desc = could not find container \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": container with ID starting with f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229120 4837 scope.go:117] "RemoveContainer" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: E1001 07:26:30.229391 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": container with ID starting with ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7 not found: ID does not exist" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229429 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} err="failed to get container status \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": rpc error: code = NotFound desc = could not find container \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": container with ID starting with ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229457 4837 scope.go:117] "RemoveContainer" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229655 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} err="failed to get container status \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": rpc error: code = NotFound desc = could not find container \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": container with ID starting with d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.229679 4837 scope.go:117] "RemoveContainer" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230271 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} err="failed to get container status \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": rpc error: code = NotFound desc = could not find container \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": container with ID starting with 367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230289 4837 scope.go:117] "RemoveContainer" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230688 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} err="failed to get container status \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": rpc error: code = NotFound desc = could not find container \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": container with ID starting with f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230723 4837 scope.go:117] "RemoveContainer" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230923 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} err="failed to get container status \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": rpc error: code = NotFound desc = could not find container \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": container with ID starting with ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.230982 4837 scope.go:117] "RemoveContainer" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231192 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} err="failed to get container status \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": rpc error: code = NotFound desc = could not find container \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": container with ID starting with d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231261 4837 scope.go:117] "RemoveContainer" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231722 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} err="failed to get container status \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": rpc error: code = NotFound desc = could not find container \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": container with ID starting with 367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231745 4837 scope.go:117] "RemoveContainer" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231977 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} err="failed to get container status \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": rpc error: code = NotFound desc = could not find container \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": container with ID starting with f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.231998 4837 scope.go:117] "RemoveContainer" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.232299 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} err="failed to get container status \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": rpc error: code = NotFound desc = could not find container \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": container with ID starting with ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.232349 4837 scope.go:117] "RemoveContainer" containerID="d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.232902 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e"} err="failed to get container status \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": rpc error: code = NotFound desc = could not find container \"d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e\": container with ID starting with d3182a82ed1e055db17b9544456fe0937ee654f3cec81c6107a5fafe627f258e not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.232962 4837 scope.go:117] "RemoveContainer" containerID="367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.233531 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58"} err="failed to get container status \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": rpc error: code = NotFound desc = could not find container \"367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58\": container with ID starting with 367421691bcb8e20012ca98446030d34026293baedccaa064bfe007d81286a58 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.233550 4837 scope.go:117] "RemoveContainer" containerID="f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.233898 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d"} err="failed to get container status \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": rpc error: code = NotFound desc = could not find container \"f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d\": container with ID starting with f47e9dd05e9c68b6d28a05c729a235c3e256577822d23bf0b873d8281dc1645d not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.233919 4837 scope.go:117] "RemoveContainer" containerID="ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.234093 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7"} err="failed to get container status \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": rpc error: code = NotFound desc = could not find container \"ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7\": container with ID starting with ea1de53adb7bbd357af7f8f0e75b2025d619a835b00fe7fae668ed6b466f4fc7 not found: ID does not exist" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.250835 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22b5q\" (UniqueName: \"kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.250931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.250977 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.251002 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.251025 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.251053 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.251073 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.258514 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: W1001 07:26:30.262735 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d64002d_6a47_4928_9495_91a58f4b2a5c.slice/crio-87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248 WatchSource:0}: Error finding container 87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248: Status 404 returned error can't find the container with id 87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248 Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352201 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352275 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352304 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352326 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352352 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352367 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352414 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22b5q\" (UniqueName: \"kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.352589 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.353005 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.357027 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.358347 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.358359 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.358683 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.368249 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22b5q\" (UniqueName: \"kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q\") pod \"ceilometer-0\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.483495 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.638739 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:30 crc kubenswrapper[4837]: I1001 07:26:30.989413 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:31 crc kubenswrapper[4837]: W1001 07:26:31.004000 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66df62a0_3e00_46c5_938c_d1a808436f7a.slice/crio-153226dca431a3942ace946b9627572f8afee3963fe33b907f843be864971f48 WatchSource:0}: Error finding container 153226dca431a3942ace946b9627572f8afee3963fe33b907f843be864971f48: Status 404 returned error can't find the container with id 153226dca431a3942ace946b9627572f8afee3963fe33b907f843be864971f48 Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.091783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerStarted","Data":"c5de51b5f6c2a76c7e0d181f7cdfb7392fdc8323c0f38c8d9eaf84c7dfe54255"} Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.091824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerStarted","Data":"87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248"} Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.097289 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerStarted","Data":"153226dca431a3942ace946b9627572f8afee3963fe33b907f843be864971f48"} Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.562274 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.621525 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.621766 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85544dfb8d-wtz4w" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-api" containerID="cri-o://93da5dd64723a812493ca4183a35399159dcfacc340b861719f66b99aa3185a3" gracePeriod=30 Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.621883 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85544dfb8d-wtz4w" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-httpd" containerID="cri-o://dd9dba269473c23311be1d264a0a8a5a066fb07a7e2e6a62e54478bad4794813" gracePeriod=30 Oct 01 07:26:31 crc kubenswrapper[4837]: I1001 07:26:31.832942 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e96695ac-c874-45ef-b080-45a0adce834f" path="/var/lib/kubelet/pods/e96695ac-c874-45ef-b080-45a0adce834f/volumes" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.107214 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerStarted","Data":"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51"} Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.110436 4837 generic.go:334] "Generic (PLEG): container finished" podID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerID="dd9dba269473c23311be1d264a0a8a5a066fb07a7e2e6a62e54478bad4794813" exitCode=0 Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.110495 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerDied","Data":"dd9dba269473c23311be1d264a0a8a5a066fb07a7e2e6a62e54478bad4794813"} Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.112458 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerStarted","Data":"e8db386829c99e752736e8a3a56fdc0ddce5b550d7fd78fdcfb1d074bd8ca628"} Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.112670 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.129729 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.129698082 podStartE2EDuration="3.129698082s" podCreationTimestamp="2025-10-01 07:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:32.128902471 +0000 UTC m=+1248.970509926" watchObservedRunningTime="2025-10-01 07:26:32.129698082 +0000 UTC m=+1248.971305537" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.527683 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2w99m"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.528839 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.541034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2w99m"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.590836 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7vb2\" (UniqueName: \"kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2\") pod \"nova-api-db-create-2w99m\" (UID: \"11fa0495-6203-43f0-aacb-534dcb9a4b04\") " pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.640845 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-skz66"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.641950 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.660186 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-skz66"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.692093 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7vb2\" (UniqueName: \"kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2\") pod \"nova-api-db-create-2w99m\" (UID: \"11fa0495-6203-43f0-aacb-534dcb9a4b04\") " pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.692137 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbptv\" (UniqueName: \"kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv\") pod \"nova-cell0-db-create-skz66\" (UID: \"eba7e445-3312-44a0-ab5e-7cdd14676b93\") " pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.739364 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7vb2\" (UniqueName: \"kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2\") pod \"nova-api-db-create-2w99m\" (UID: \"11fa0495-6203-43f0-aacb-534dcb9a4b04\") " pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.743609 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jd57t"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.745215 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.758733 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jd57t"] Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.795976 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlwx5\" (UniqueName: \"kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5\") pod \"nova-cell1-db-create-jd57t\" (UID: \"4420e1fc-5656-453c-b5a4-31d5d7fe0012\") " pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.796083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbptv\" (UniqueName: \"kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv\") pod \"nova-cell0-db-create-skz66\" (UID: \"eba7e445-3312-44a0-ab5e-7cdd14676b93\") " pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.811952 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbptv\" (UniqueName: \"kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv\") pod \"nova-cell0-db-create-skz66\" (UID: \"eba7e445-3312-44a0-ab5e-7cdd14676b93\") " pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.852891 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.897951 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlwx5\" (UniqueName: \"kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5\") pod \"nova-cell1-db-create-jd57t\" (UID: \"4420e1fc-5656-453c-b5a4-31d5d7fe0012\") " pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.924547 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlwx5\" (UniqueName: \"kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5\") pod \"nova-cell1-db-create-jd57t\" (UID: \"4420e1fc-5656-453c-b5a4-31d5d7fe0012\") " pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.956991 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 07:26:32 crc kubenswrapper[4837]: I1001 07:26:32.985104 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.148827 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.150242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerStarted","Data":"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca"} Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.209877 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.210206 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.210436 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="dnsmasq-dns" containerID="cri-o://8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df" gracePeriod=10 Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.313839 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.357883 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:33 crc kubenswrapper[4837]: W1001 07:26:33.657811 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11fa0495_6203_43f0_aacb_534dcb9a4b04.slice/crio-96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7 WatchSource:0}: Error finding container 96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7: Status 404 returned error can't find the container with id 96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7 Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.664300 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: connect: connection refused" Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.668758 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-skz66"] Oct 01 07:26:33 crc kubenswrapper[4837]: I1001 07:26:33.689267 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2w99m"] Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.025689 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jd57t"] Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.089196 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.180204 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerStarted","Data":"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.195229 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2w99m" event={"ID":"11fa0495-6203-43f0-aacb-534dcb9a4b04","Type":"ContainerStarted","Data":"96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.208452 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jd57t" event={"ID":"4420e1fc-5656-453c-b5a4-31d5d7fe0012","Type":"ContainerStarted","Data":"e0f9a36a211e98945f9f5206186791ed02dafaeb184e040c14f2c2901a74f8a2"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.215115 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-skz66" event={"ID":"eba7e445-3312-44a0-ab5e-7cdd14676b93","Type":"ContainerStarted","Data":"fdeb71818cbf9a241d4aebcb770242571e40d8e9d3744310144a88e43c05d2e7"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229334 4837 generic.go:334] "Generic (PLEG): container finished" podID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerID="8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df" exitCode=0 Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229454 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229506 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" event={"ID":"b5f90d53-0db2-4d86-8bda-8ff78e80c604","Type":"ContainerDied","Data":"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229530 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c964fd46f-pbtxw" event={"ID":"b5f90d53-0db2-4d86-8bda-8ff78e80c604","Type":"ContainerDied","Data":"cf731f0bb246b88e81017a974579d3ae7108ca6fb9be06051ac979cacdf90bd8"} Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229526 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="cinder-scheduler" containerID="cri-o://ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc" gracePeriod=30 Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229545 4837 scope.go:117] "RemoveContainer" containerID="8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.229707 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="probe" containerID="cri-o://6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290" gracePeriod=30 Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.234728 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-skz66" podStartSLOduration=2.234699584 podStartE2EDuration="2.234699584s" podCreationTimestamp="2025-10-01 07:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:34.233804743 +0000 UTC m=+1251.075412218" watchObservedRunningTime="2025-10-01 07:26:34.234699584 +0000 UTC m=+1251.076307039" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247209 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247332 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247352 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247397 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247457 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb4qx\" (UniqueName: \"kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.247561 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb\") pod \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\" (UID: \"b5f90d53-0db2-4d86-8bda-8ff78e80c604\") " Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.259781 4837 scope.go:117] "RemoveContainer" containerID="a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.282826 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx" (OuterVolumeSpecName: "kube-api-access-pb4qx") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "kube-api-access-pb4qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.300681 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.307970 4837 scope.go:117] "RemoveContainer" containerID="8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df" Oct 01 07:26:34 crc kubenswrapper[4837]: E1001 07:26:34.310353 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df\": container with ID starting with 8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df not found: ID does not exist" containerID="8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.310407 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df"} err="failed to get container status \"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df\": rpc error: code = NotFound desc = could not find container \"8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df\": container with ID starting with 8bd576c89e4e73620f897cb4ef306b0340b260285fd26f5328101a8e323be6df not found: ID does not exist" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.310435 4837 scope.go:117] "RemoveContainer" containerID="a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020" Oct 01 07:26:34 crc kubenswrapper[4837]: E1001 07:26:34.310804 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020\": container with ID starting with a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020 not found: ID does not exist" containerID="a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.310894 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020"} err="failed to get container status \"a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020\": rpc error: code = NotFound desc = could not find container \"a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020\": container with ID starting with a83901673dceba078cfbcda489c93003d647a33c120a2384e91465cb2df88020 not found: ID does not exist" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.332334 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.336134 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.338588 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.340959 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config" (OuterVolumeSpecName: "config") pod "b5f90d53-0db2-4d86-8bda-8ff78e80c604" (UID: "b5f90d53-0db2-4d86-8bda-8ff78e80c604"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349114 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349152 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349162 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349173 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb4qx\" (UniqueName: \"kubernetes.io/projected/b5f90d53-0db2-4d86-8bda-8ff78e80c604-kube-api-access-pb4qx\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349183 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.349191 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f90d53-0db2-4d86-8bda-8ff78e80c604-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.573343 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:26:34 crc kubenswrapper[4837]: I1001 07:26:34.582237 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c964fd46f-pbtxw"] Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.183503 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.279845 4837 generic.go:334] "Generic (PLEG): container finished" podID="1152254c-511c-40e0-8147-b607c72cddf3" containerID="6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290" exitCode=0 Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.280085 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerDied","Data":"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290"} Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.291830 4837 generic.go:334] "Generic (PLEG): container finished" podID="11fa0495-6203-43f0-aacb-534dcb9a4b04" containerID="8453b03acddb5b9c4dbf1558763948d0e396c1231e9e3ea725a0e06c4ac8956d" exitCode=0 Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.292620 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2w99m" event={"ID":"11fa0495-6203-43f0-aacb-534dcb9a4b04","Type":"ContainerDied","Data":"8453b03acddb5b9c4dbf1558763948d0e396c1231e9e3ea725a0e06c4ac8956d"} Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.306021 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.308371 4837 generic.go:334] "Generic (PLEG): container finished" podID="4420e1fc-5656-453c-b5a4-31d5d7fe0012" containerID="d863c9fe0ae9e19204566c6cb4300be788afaf2105039c052930acb1e3a11e23" exitCode=0 Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.308434 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jd57t" event={"ID":"4420e1fc-5656-453c-b5a4-31d5d7fe0012","Type":"ContainerDied","Data":"d863c9fe0ae9e19204566c6cb4300be788afaf2105039c052930acb1e3a11e23"} Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.313064 4837 generic.go:334] "Generic (PLEG): container finished" podID="eba7e445-3312-44a0-ab5e-7cdd14676b93" containerID="cdd54b9fd4aba5b9af8e7129caeebcabaf6de2161132e3fe487197e5759bae78" exitCode=0 Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.313111 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-skz66" event={"ID":"eba7e445-3312-44a0-ab5e-7cdd14676b93","Type":"ContainerDied","Data":"cdd54b9fd4aba5b9af8e7129caeebcabaf6de2161132e3fe487197e5759bae78"} Oct 01 07:26:35 crc kubenswrapper[4837]: I1001 07:26:35.826694 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" path="/var/lib/kubelet/pods/b5f90d53-0db2-4d86-8bda-8ff78e80c604/volumes" Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.337066 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerStarted","Data":"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3"} Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.337424 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-central-agent" containerID="cri-o://6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51" gracePeriod=30 Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.337964 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="proxy-httpd" containerID="cri-o://733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3" gracePeriod=30 Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.338014 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="sg-core" containerID="cri-o://18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0" gracePeriod=30 Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.338050 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-notification-agent" containerID="cri-o://093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca" gracePeriod=30 Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.370423 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.955032363 podStartE2EDuration="6.370406668s" podCreationTimestamp="2025-10-01 07:26:30 +0000 UTC" firstStartedPulling="2025-10-01 07:26:31.006641075 +0000 UTC m=+1247.848248530" lastFinishedPulling="2025-10-01 07:26:35.42201538 +0000 UTC m=+1252.263622835" observedRunningTime="2025-10-01 07:26:36.369683249 +0000 UTC m=+1253.211290724" watchObservedRunningTime="2025-10-01 07:26:36.370406668 +0000 UTC m=+1253.212014123" Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.826778 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.875259 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.882168 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.904521 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlwx5\" (UniqueName: \"kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5\") pod \"4420e1fc-5656-453c-b5a4-31d5d7fe0012\" (UID: \"4420e1fc-5656-453c-b5a4-31d5d7fe0012\") " Oct 01 07:26:36 crc kubenswrapper[4837]: I1001 07:26:36.910109 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5" (OuterVolumeSpecName: "kube-api-access-wlwx5") pod "4420e1fc-5656-453c-b5a4-31d5d7fe0012" (UID: "4420e1fc-5656-453c-b5a4-31d5d7fe0012"). InnerVolumeSpecName "kube-api-access-wlwx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.006190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7vb2\" (UniqueName: \"kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2\") pod \"11fa0495-6203-43f0-aacb-534dcb9a4b04\" (UID: \"11fa0495-6203-43f0-aacb-534dcb9a4b04\") " Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.006399 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbptv\" (UniqueName: \"kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv\") pod \"eba7e445-3312-44a0-ab5e-7cdd14676b93\" (UID: \"eba7e445-3312-44a0-ab5e-7cdd14676b93\") " Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.006943 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlwx5\" (UniqueName: \"kubernetes.io/projected/4420e1fc-5656-453c-b5a4-31d5d7fe0012-kube-api-access-wlwx5\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.010877 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2" (OuterVolumeSpecName: "kube-api-access-q7vb2") pod "11fa0495-6203-43f0-aacb-534dcb9a4b04" (UID: "11fa0495-6203-43f0-aacb-534dcb9a4b04"). InnerVolumeSpecName "kube-api-access-q7vb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.010976 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv" (OuterVolumeSpecName: "kube-api-access-gbptv") pod "eba7e445-3312-44a0-ab5e-7cdd14676b93" (UID: "eba7e445-3312-44a0-ab5e-7cdd14676b93"). InnerVolumeSpecName "kube-api-access-gbptv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.108398 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbptv\" (UniqueName: \"kubernetes.io/projected/eba7e445-3312-44a0-ab5e-7cdd14676b93-kube-api-access-gbptv\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.108427 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7vb2\" (UniqueName: \"kubernetes.io/projected/11fa0495-6203-43f0-aacb-534dcb9a4b04-kube-api-access-q7vb2\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364789 4837 generic.go:334] "Generic (PLEG): container finished" podID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerID="733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3" exitCode=0 Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364836 4837 generic.go:334] "Generic (PLEG): container finished" podID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerID="18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0" exitCode=2 Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364852 4837 generic.go:334] "Generic (PLEG): container finished" podID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerID="093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca" exitCode=0 Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364845 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerDied","Data":"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364963 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerDied","Data":"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.364976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerDied","Data":"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.366499 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2w99m" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.366505 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2w99m" event={"ID":"11fa0495-6203-43f0-aacb-534dcb9a4b04","Type":"ContainerDied","Data":"96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.366547 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96a9253de3bd022a53f48ca25bf7407842f4d5ce6452e53c8f7e5d4f993ca0b7" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.369239 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jd57t" event={"ID":"4420e1fc-5656-453c-b5a4-31d5d7fe0012","Type":"ContainerDied","Data":"e0f9a36a211e98945f9f5206186791ed02dafaeb184e040c14f2c2901a74f8a2"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.369269 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f9a36a211e98945f9f5206186791ed02dafaeb184e040c14f2c2901a74f8a2" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.369350 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jd57t" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.372552 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-skz66" event={"ID":"eba7e445-3312-44a0-ab5e-7cdd14676b93","Type":"ContainerDied","Data":"fdeb71818cbf9a241d4aebcb770242571e40d8e9d3744310144a88e43c05d2e7"} Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.372572 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdeb71818cbf9a241d4aebcb770242571e40d8e9d3744310144a88e43c05d2e7" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.372643 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-skz66" Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.885495 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.886021 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-log" containerID="cri-o://a2a36e3ba660fceacc1d7e98c927ee4199206baa6f52ac1dced189d0a31882c6" gracePeriod=30 Oct 01 07:26:37 crc kubenswrapper[4837]: I1001 07:26:37.886374 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-httpd" containerID="cri-o://db3da1ff91657924b3e2d7ae07a63ed06b13f469b3bbe00f8607fc6501c9abd5" gracePeriod=30 Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.382944 4837 generic.go:334] "Generic (PLEG): container finished" podID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerID="a2a36e3ba660fceacc1d7e98c927ee4199206baa6f52ac1dced189d0a31882c6" exitCode=143 Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.383132 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerDied","Data":"a2a36e3ba660fceacc1d7e98c927ee4199206baa6f52ac1dced189d0a31882c6"} Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.385094 4837 generic.go:334] "Generic (PLEG): container finished" podID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerID="93da5dd64723a812493ca4183a35399159dcfacc340b861719f66b99aa3185a3" exitCode=0 Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.385133 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerDied","Data":"93da5dd64723a812493ca4183a35399159dcfacc340b861719f66b99aa3185a3"} Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.385160 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85544dfb8d-wtz4w" event={"ID":"c28f68bd-47d2-4534-8e1d-499073eed9f6","Type":"ContainerDied","Data":"a4459cbf7d3ef13f30ca7ca41949068142126760e65a776d9a5c34f25671ebd1"} Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.385171 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4459cbf7d3ef13f30ca7ca41949068142126760e65a776d9a5c34f25671ebd1" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.451797 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.531619 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle\") pod \"c28f68bd-47d2-4534-8e1d-499073eed9f6\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.531851 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config\") pod \"c28f68bd-47d2-4534-8e1d-499073eed9f6\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.531879 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8v9p\" (UniqueName: \"kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p\") pod \"c28f68bd-47d2-4534-8e1d-499073eed9f6\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.531898 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config\") pod \"c28f68bd-47d2-4534-8e1d-499073eed9f6\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.531946 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs\") pod \"c28f68bd-47d2-4534-8e1d-499073eed9f6\" (UID: \"c28f68bd-47d2-4534-8e1d-499073eed9f6\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.542300 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p" (OuterVolumeSpecName: "kube-api-access-k8v9p") pod "c28f68bd-47d2-4534-8e1d-499073eed9f6" (UID: "c28f68bd-47d2-4534-8e1d-499073eed9f6"). InnerVolumeSpecName "kube-api-access-k8v9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.543018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c28f68bd-47d2-4534-8e1d-499073eed9f6" (UID: "c28f68bd-47d2-4534-8e1d-499073eed9f6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.608055 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c28f68bd-47d2-4534-8e1d-499073eed9f6" (UID: "c28f68bd-47d2-4534-8e1d-499073eed9f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.628495 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config" (OuterVolumeSpecName: "config") pod "c28f68bd-47d2-4534-8e1d-499073eed9f6" (UID: "c28f68bd-47d2-4534-8e1d-499073eed9f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.633757 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.633795 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8v9p\" (UniqueName: \"kubernetes.io/projected/c28f68bd-47d2-4534-8e1d-499073eed9f6-kube-api-access-k8v9p\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.633807 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.633815 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.678371 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c28f68bd-47d2-4534-8e1d-499073eed9f6" (UID: "c28f68bd-47d2-4534-8e1d-499073eed9f6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.736795 4837 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c28f68bd-47d2-4534-8e1d-499073eed9f6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.794539 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.935623 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939239 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22b5q\" (UniqueName: \"kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939392 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939432 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939510 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939544 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939568 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts\") pod \"66df62a0-3e00-46c5-938c-d1a808436f7a\" (UID: \"66df62a0-3e00-46c5-938c-d1a808436f7a\") " Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.939875 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.940168 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.940590 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-log" containerID="cri-o://cc574cbffee90ee1ee641c1741dcc43c1f2b4b2221b367eaec963ac066577b55" gracePeriod=30 Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.941039 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-httpd" containerID="cri-o://9a6a1a98f80d99ded752635627bf4357fc7c57b99d84065d9f695f84bd4c75e9" gracePeriod=30 Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.943415 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q" (OuterVolumeSpecName: "kube-api-access-22b5q") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "kube-api-access-22b5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.944560 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts" (OuterVolumeSpecName: "scripts") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.950854 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.950885 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66df62a0-3e00-46c5-938c-d1a808436f7a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:38 crc kubenswrapper[4837]: I1001 07:26:38.997788 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.052730 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.052918 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.052984 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22b5q\" (UniqueName: \"kubernetes.io/projected/66df62a0-3e00-46c5-938c-d1a808436f7a-kube-api-access-22b5q\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.091822 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.097833 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data" (OuterVolumeSpecName: "config-data") pod "66df62a0-3e00-46c5-938c-d1a808436f7a" (UID: "66df62a0-3e00-46c5-938c-d1a808436f7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.154055 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.154218 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66df62a0-3e00-46c5-938c-d1a808436f7a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.155361 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255434 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42k4j\" (UniqueName: \"kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255513 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255560 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255596 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255649 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle\") pod \"1152254c-511c-40e0-8147-b607c72cddf3\" (UID: \"1152254c-511c-40e0-8147-b607c72cddf3\") " Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.255881 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.256685 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1152254c-511c-40e0-8147-b607c72cddf3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.263833 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j" (OuterVolumeSpecName: "kube-api-access-42k4j") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "kube-api-access-42k4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.264067 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.264205 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts" (OuterVolumeSpecName: "scripts") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.315889 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.360216 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.360247 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.360256 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.360264 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42k4j\" (UniqueName: \"kubernetes.io/projected/1152254c-511c-40e0-8147-b607c72cddf3-kube-api-access-42k4j\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.374221 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data" (OuterVolumeSpecName: "config-data") pod "1152254c-511c-40e0-8147-b607c72cddf3" (UID: "1152254c-511c-40e0-8147-b607c72cddf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.395774 4837 generic.go:334] "Generic (PLEG): container finished" podID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerID="6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51" exitCode=0 Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.395831 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerDied","Data":"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51"} Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.395857 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66df62a0-3e00-46c5-938c-d1a808436f7a","Type":"ContainerDied","Data":"153226dca431a3942ace946b9627572f8afee3963fe33b907f843be864971f48"} Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.395876 4837 scope.go:117] "RemoveContainer" containerID="733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.396009 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.407058 4837 generic.go:334] "Generic (PLEG): container finished" podID="1152254c-511c-40e0-8147-b607c72cddf3" containerID="ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc" exitCode=0 Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.407182 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.407140 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerDied","Data":"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc"} Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.407512 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1152254c-511c-40e0-8147-b607c72cddf3","Type":"ContainerDied","Data":"261b354f4530c354587c143c410bad9980c3e53191e0942f04fc35787242aa02"} Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.411404 4837 generic.go:334] "Generic (PLEG): container finished" podID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerID="cc574cbffee90ee1ee641c1741dcc43c1f2b4b2221b367eaec963ac066577b55" exitCode=143 Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.411483 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85544dfb8d-wtz4w" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.411825 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerDied","Data":"cc574cbffee90ee1ee641c1741dcc43c1f2b4b2221b367eaec963ac066577b55"} Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.421456 4837 scope.go:117] "RemoveContainer" containerID="18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.448034 4837 scope.go:117] "RemoveContainer" containerID="093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.448839 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.457490 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.461669 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152254c-511c-40e0-8147-b607c72cddf3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.474129 4837 scope.go:117] "RemoveContainer" containerID="6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.491819 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.520385 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539054 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539540 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="init" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539560 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="init" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539572 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="cinder-scheduler" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539578 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="cinder-scheduler" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539590 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-notification-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539602 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-notification-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539609 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4420e1fc-5656-453c-b5a4-31d5d7fe0012" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539617 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4420e1fc-5656-453c-b5a4-31d5d7fe0012" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539628 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba7e445-3312-44a0-ab5e-7cdd14676b93" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539633 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba7e445-3312-44a0-ab5e-7cdd14676b93" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539643 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="proxy-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539649 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="proxy-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539661 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-central-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539667 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-central-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539678 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="dnsmasq-dns" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539685 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="dnsmasq-dns" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539712 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-api" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539718 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-api" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539729 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539734 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539746 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="probe" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539752 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="probe" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539768 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fa0495-6203-43f0-aacb-534dcb9a4b04" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539774 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fa0495-6203-43f0-aacb-534dcb9a4b04" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.539785 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="sg-core" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539791 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="sg-core" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539950 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="sg-core" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539963 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="cinder-scheduler" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539972 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-notification-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539982 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba7e445-3312-44a0-ab5e-7cdd14676b93" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.539992 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-api" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540002 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="proxy-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540012 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fa0495-6203-43f0-aacb-534dcb9a4b04" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540026 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f90d53-0db2-4d86-8bda-8ff78e80c604" containerName="dnsmasq-dns" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540036 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4420e1fc-5656-453c-b5a4-31d5d7fe0012" containerName="mariadb-database-create" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540045 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1152254c-511c-40e0-8147-b607c72cddf3" containerName="probe" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540053 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" containerName="neutron-httpd" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.540066 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" containerName="ceilometer-central-agent" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.542389 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.546345 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.552161 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.563180 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.568611 4837 scope.go:117] "RemoveContainer" containerID="733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.569083 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3\": container with ID starting with 733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3 not found: ID does not exist" containerID="733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569122 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3"} err="failed to get container status \"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3\": rpc error: code = NotFound desc = could not find container \"733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3\": container with ID starting with 733e2825a6bccf49ebcf63e33f2de8bc224e488ae7d5d6695f57e5ad3bcdb6b3 not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569170 4837 scope.go:117] "RemoveContainer" containerID="18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.569674 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0\": container with ID starting with 18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0 not found: ID does not exist" containerID="18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569713 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0"} err="failed to get container status \"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0\": rpc error: code = NotFound desc = could not find container \"18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0\": container with ID starting with 18590f7f4b42cb50664bd43907db87a00d62dc9f6114ee0d0c7ee540af5e62a0 not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569736 4837 scope.go:117] "RemoveContainer" containerID="093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.569909 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca\": container with ID starting with 093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca not found: ID does not exist" containerID="093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569924 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca"} err="failed to get container status \"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca\": rpc error: code = NotFound desc = could not find container \"093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca\": container with ID starting with 093f5b20280e5fcd3617b38f0775f1e22b60527a9d51c695182b0331b7be53ca not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.569937 4837 scope.go:117] "RemoveContainer" containerID="6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.570101 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51\": container with ID starting with 6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51 not found: ID does not exist" containerID="6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.570116 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51"} err="failed to get container status \"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51\": rpc error: code = NotFound desc = could not find container \"6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51\": container with ID starting with 6905fef3716eace7f880e32c2503d923bc3e09bd675ea1690fdeada11e3cad51 not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.570128 4837 scope.go:117] "RemoveContainer" containerID="6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.573918 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85544dfb8d-wtz4w"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.582062 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.585328 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.590122 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.590489 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.598375 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.605615 4837 scope.go:117] "RemoveContainer" containerID="ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.634409 4837 scope.go:117] "RemoveContainer" containerID="6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.634943 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290\": container with ID starting with 6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290 not found: ID does not exist" containerID="6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.634973 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290"} err="failed to get container status \"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290\": rpc error: code = NotFound desc = could not find container \"6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290\": container with ID starting with 6e245cddcd9276ba6c7d50b732aa98d59707bef8f64b91790a03111e0df0a290 not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.634995 4837 scope.go:117] "RemoveContainer" containerID="ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc" Oct 01 07:26:39 crc kubenswrapper[4837]: E1001 07:26:39.635312 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc\": container with ID starting with ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc not found: ID does not exist" containerID="ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.635343 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc"} err="failed to get container status \"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc\": rpc error: code = NotFound desc = could not find container \"ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc\": container with ID starting with ebfb3ac40df97294eb628ae7739fbffdec6850bf5155f2c147a926dfa33153dc not found: ID does not exist" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.665840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.665886 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9cbb\" (UniqueName: \"kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.665964 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.665983 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.666094 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.666216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.666260 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.767616 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.767985 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768005 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768024 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768055 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768084 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768102 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fm9r\" (UniqueName: \"kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768128 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768149 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768175 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768193 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768215 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9cbb\" (UniqueName: \"kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.768275 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.769299 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.771920 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.774661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.782957 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.785954 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.787408 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.790118 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9cbb\" (UniqueName: \"kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb\") pod \"ceilometer-0\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.825360 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1152254c-511c-40e0-8147-b607c72cddf3" path="/var/lib/kubelet/pods/1152254c-511c-40e0-8147-b607c72cddf3/volumes" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.825958 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66df62a0-3e00-46c5-938c-d1a808436f7a" path="/var/lib/kubelet/pods/66df62a0-3e00-46c5-938c-d1a808436f7a/volumes" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.826830 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c28f68bd-47d2-4534-8e1d-499073eed9f6" path="/var/lib/kubelet/pods/c28f68bd-47d2-4534-8e1d-499073eed9f6/volumes" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869663 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869741 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fm9r\" (UniqueName: \"kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869771 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869865 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869907 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.869946 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.870397 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.870855 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.873419 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.874157 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.875370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.878276 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.899363 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fm9r\" (UniqueName: \"kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r\") pod \"cinder-scheduler-0\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " pod="openstack/cinder-scheduler-0" Oct 01 07:26:39 crc kubenswrapper[4837]: I1001 07:26:39.912206 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.351524 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:40 crc kubenswrapper[4837]: W1001 07:26:40.355223 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e800fd_f1e2_46e2_a806_d17c54adb0a2.slice/crio-731d326c4923f1a7071c0c258cbba267ab57d637b9f99801c2d3e608a252b61c WatchSource:0}: Error finding container 731d326c4923f1a7071c0c258cbba267ab57d637b9f99801c2d3e608a252b61c: Status 404 returned error can't find the container with id 731d326c4923f1a7071c0c258cbba267ab57d637b9f99801c2d3e608a252b61c Oct 01 07:26:40 crc kubenswrapper[4837]: W1001 07:26:40.427051 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70ff50da_fcbe_4fe9_9a35_073f2abdc3e7.slice/crio-d9c89ed9bfc11964e0d5b30ae091d4dc0f4c1763318f924641f2df54697e6807 WatchSource:0}: Error finding container d9c89ed9bfc11964e0d5b30ae091d4dc0f4c1763318f924641f2df54697e6807: Status 404 returned error can't find the container with id d9c89ed9bfc11964e0d5b30ae091d4dc0f4c1763318f924641f2df54697e6807 Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.429898 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.431577 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerStarted","Data":"731d326c4923f1a7071c0c258cbba267ab57d637b9f99801c2d3e608a252b61c"} Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.615729 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.646822 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.727202 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.813882 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.814202 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b5db87b6d-sq97p" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api-log" containerID="cri-o://12ec459428014c7466214d2c99d3446b10ae3bcac80009aa60ef4f936e7fd003" gracePeriod=30 Oct 01 07:26:40 crc kubenswrapper[4837]: I1001 07:26:40.814445 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b5db87b6d-sq97p" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api" containerID="cri-o://e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa" gracePeriod=30 Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.454498 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerStarted","Data":"73ae4948895720b7496ecb1982bf5fbab9c712bead101852e3e795d3237de601"} Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.454916 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerStarted","Data":"d9c89ed9bfc11964e0d5b30ae091d4dc0f4c1763318f924641f2df54697e6807"} Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.470458 4837 generic.go:334] "Generic (PLEG): container finished" podID="7243da66-6008-464f-a5f0-423e342f34d0" containerID="12ec459428014c7466214d2c99d3446b10ae3bcac80009aa60ef4f936e7fd003" exitCode=143 Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.470550 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerDied","Data":"12ec459428014c7466214d2c99d3446b10ae3bcac80009aa60ef4f936e7fd003"} Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.473885 4837 generic.go:334] "Generic (PLEG): container finished" podID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerID="db3da1ff91657924b3e2d7ae07a63ed06b13f469b3bbe00f8607fc6501c9abd5" exitCode=0 Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.473924 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerDied","Data":"db3da1ff91657924b3e2d7ae07a63ed06b13f469b3bbe00f8607fc6501c9abd5"} Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.482650 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerStarted","Data":"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa"} Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.572271 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720406 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720503 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720539 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720560 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgzhh\" (UniqueName: \"kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720611 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720633 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.720672 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.721133 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle\") pod \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\" (UID: \"6bd79b05-167a-4b5c-9afe-b17058cb13bc\") " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.727460 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs" (OuterVolumeSpecName: "logs") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.727729 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.727810 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts" (OuterVolumeSpecName: "scripts") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.727833 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh" (OuterVolumeSpecName: "kube-api-access-mgzhh") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "kube-api-access-mgzhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.734860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.806095 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825664 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825729 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825741 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgzhh\" (UniqueName: \"kubernetes.io/projected/6bd79b05-167a-4b5c-9afe-b17058cb13bc-kube-api-access-mgzhh\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825751 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825759 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd79b05-167a-4b5c-9afe-b17058cb13bc-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.825767 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.847088 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.875614 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.886938 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data" (OuterVolumeSpecName: "config-data") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.901200 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6bd79b05-167a-4b5c-9afe-b17058cb13bc" (UID: "6bd79b05-167a-4b5c-9afe-b17058cb13bc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.927757 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.927799 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:41 crc kubenswrapper[4837]: I1001 07:26:41.927809 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd79b05-167a-4b5c-9afe-b17058cb13bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.496799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerStarted","Data":"bf54d4f5f766e15a07eac31a84eb955d37b5b5dfce2d9dbf7292f0ae70e3db3f"} Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.500323 4837 generic.go:334] "Generic (PLEG): container finished" podID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerID="9a6a1a98f80d99ded752635627bf4357fc7c57b99d84065d9f695f84bd4c75e9" exitCode=0 Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.500376 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerDied","Data":"9a6a1a98f80d99ded752635627bf4357fc7c57b99d84065d9f695f84bd4c75e9"} Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.502645 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd79b05-167a-4b5c-9afe-b17058cb13bc","Type":"ContainerDied","Data":"f140d4a2427401013beea5445e820c81d1b8a2720cc40ac5d452f52df83146d7"} Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.502673 4837 scope.go:117] "RemoveContainer" containerID="db3da1ff91657924b3e2d7ae07a63ed06b13f469b3bbe00f8607fc6501c9abd5" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.502797 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.514522 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerStarted","Data":"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af"} Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.546479 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.54645852 podStartE2EDuration="3.54645852s" podCreationTimestamp="2025-10-01 07:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:42.526121167 +0000 UTC m=+1259.367728622" watchObservedRunningTime="2025-10-01 07:26:42.54645852 +0000 UTC m=+1259.388065965" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.638828 4837 scope.go:117] "RemoveContainer" containerID="a2a36e3ba660fceacc1d7e98c927ee4199206baa6f52ac1dced189d0a31882c6" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.662400 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.674440 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.680488 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.696784 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:42 crc kubenswrapper[4837]: E1001 07:26:42.697212 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697231 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: E1001 07:26:42.697245 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697252 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: E1001 07:26:42.697275 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697282 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: E1001 07:26:42.697298 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697304 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697483 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697513 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697530 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-log" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.697543 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" containerName="glance-httpd" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.698512 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.703601 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.703765 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.764044 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848144 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848382 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848452 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62j52\" (UniqueName: \"kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848565 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848719 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848786 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848851 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.848928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs\") pod \"a9fcd80f-2947-4f8a-aa50-4b4006035393\" (UID: \"a9fcd80f-2947-4f8a-aa50-4b4006035393\") " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849041 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849238 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs" (OuterVolumeSpecName: "logs") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849345 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849509 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849726 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m68mq\" (UniqueName: \"kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.849919 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.850092 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.850138 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.850409 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.850453 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9fcd80f-2947-4f8a-aa50-4b4006035393-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.852272 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52" (OuterVolumeSpecName: "kube-api-access-62j52") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "kube-api-access-62j52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.853842 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.863277 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts" (OuterVolumeSpecName: "scripts") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.878014 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.911797 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data" (OuterVolumeSpecName: "config-data") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.941805 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a9fcd80f-2947-4f8a-aa50-4b4006035393" (UID: "a9fcd80f-2947-4f8a-aa50-4b4006035393"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952136 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952193 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952226 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952267 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952327 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m68mq\" (UniqueName: \"kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952348 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952384 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952399 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952469 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952480 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62j52\" (UniqueName: \"kubernetes.io/projected/a9fcd80f-2947-4f8a-aa50-4b4006035393-kube-api-access-62j52\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952488 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952497 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952514 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952524 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9fcd80f-2947-4f8a-aa50-4b4006035393-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.952537 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.953347 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.955624 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.955899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.957178 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.958979 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.961671 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.972559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m68mq\" (UniqueName: \"kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.978886 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 01 07:26:42 crc kubenswrapper[4837]: I1001 07:26:42.987244 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " pod="openstack/glance-default-external-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.018991 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.053709 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.527615 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.527628 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a9fcd80f-2947-4f8a-aa50-4b4006035393","Type":"ContainerDied","Data":"4f77b3c3d1afee943b4cc1259496e89befab677fcc2ebe167d5561a153e61d31"} Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.528194 4837 scope.go:117] "RemoveContainer" containerID="9a6a1a98f80d99ded752635627bf4357fc7c57b99d84065d9f695f84bd4c75e9" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.535740 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.540141 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerStarted","Data":"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1"} Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.556514 4837 scope.go:117] "RemoveContainer" containerID="cc574cbffee90ee1ee641c1741dcc43c1f2b4b2221b367eaec963ac066577b55" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.622308 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.649744 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.667744 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.669264 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.674130 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.675007 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.675674 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768136 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768219 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768251 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768306 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm4sj\" (UniqueName: \"kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.768368 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.829219 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bd79b05-167a-4b5c-9afe-b17058cb13bc" path="/var/lib/kubelet/pods/6bd79b05-167a-4b5c-9afe-b17058cb13bc/volumes" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.830537 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9fcd80f-2947-4f8a-aa50-4b4006035393" path="/var/lib/kubelet/pods/a9fcd80f-2947-4f8a-aa50-4b4006035393/volumes" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.869825 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.869868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm4sj\" (UniqueName: \"kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.869935 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870003 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870023 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870079 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870097 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870326 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870494 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.870794 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.875167 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.875374 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.878459 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.884263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.890686 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.893729 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.893964 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm4sj\" (UniqueName: \"kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.901940 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.988240 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.991387 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b5db87b6d-sq97p" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:34180->10.217.0.162:9311: read: connection reset by peer" Oct 01 07:26:43 crc kubenswrapper[4837]: I1001 07:26:43.991540 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b5db87b6d-sq97p" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:34166->10.217.0.162:9311: read: connection reset by peer" Oct 01 07:26:44 crc kubenswrapper[4837]: E1001 07:26:44.261427 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7243da66_6008_464f_a5f0_423e342f34d0.slice/crio-conmon-e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7243da66_6008_464f_a5f0_423e342f34d0.slice/crio-e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.567735 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerStarted","Data":"4db0657ecb1d6bab9cfa0aca6d6c4da781d4dc1788c9788ff50b8ac94c614f0d"} Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.568079 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerStarted","Data":"c28dbbd5d84b3f270ac6edb06187328bd34807ae9cf8ba2af98e74e863739a57"} Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.577588 4837 generic.go:334] "Generic (PLEG): container finished" podID="7243da66-6008-464f-a5f0-423e342f34d0" containerID="e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa" exitCode=0 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.577647 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerDied","Data":"e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa"} Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.577667 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b5db87b6d-sq97p" event={"ID":"7243da66-6008-464f-a5f0-423e342f34d0","Type":"ContainerDied","Data":"dd92804295a499b6034fc83afb8968bbde241ba43cbace5d43dac9b0edf210b2"} Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.577680 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd92804295a499b6034fc83afb8968bbde241ba43cbace5d43dac9b0edf210b2" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerStarted","Data":"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654"} Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586209 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-central-agent" containerID="cri-o://e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa" gracePeriod=30 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586282 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="sg-core" containerID="cri-o://477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1" gracePeriod=30 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586301 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="proxy-httpd" containerID="cri-o://b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654" gracePeriod=30 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586340 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.586324 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-notification-agent" containerID="cri-o://e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af" gracePeriod=30 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.588226 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.612741 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.72016124 podStartE2EDuration="5.612726085s" podCreationTimestamp="2025-10-01 07:26:39 +0000 UTC" firstStartedPulling="2025-10-01 07:26:40.35748468 +0000 UTC m=+1257.199092135" lastFinishedPulling="2025-10-01 07:26:44.250049525 +0000 UTC m=+1261.091656980" observedRunningTime="2025-10-01 07:26:44.603825505 +0000 UTC m=+1261.445432960" watchObservedRunningTime="2025-10-01 07:26:44.612726085 +0000 UTC m=+1261.454333540" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.689402 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data\") pod \"7243da66-6008-464f-a5f0-423e342f34d0\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.689454 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle\") pod \"7243da66-6008-464f-a5f0-423e342f34d0\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.689522 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs\") pod \"7243da66-6008-464f-a5f0-423e342f34d0\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.689616 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wglnl\" (UniqueName: \"kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl\") pod \"7243da66-6008-464f-a5f0-423e342f34d0\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.689668 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom\") pod \"7243da66-6008-464f-a5f0-423e342f34d0\" (UID: \"7243da66-6008-464f-a5f0-423e342f34d0\") " Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.690318 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs" (OuterVolumeSpecName: "logs") pod "7243da66-6008-464f-a5f0-423e342f34d0" (UID: "7243da66-6008-464f-a5f0-423e342f34d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.694121 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7243da66-6008-464f-a5f0-423e342f34d0" (UID: "7243da66-6008-464f-a5f0-423e342f34d0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.695126 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl" (OuterVolumeSpecName: "kube-api-access-wglnl") pod "7243da66-6008-464f-a5f0-423e342f34d0" (UID: "7243da66-6008-464f-a5f0-423e342f34d0"). InnerVolumeSpecName "kube-api-access-wglnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.736770 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7243da66-6008-464f-a5f0-423e342f34d0" (UID: "7243da66-6008-464f-a5f0-423e342f34d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.758482 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:26:44 crc kubenswrapper[4837]: W1001 07:26:44.766994 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73e65e7c_ced2_4e38_9841_3f383471b275.slice/crio-57b012ebf793a59f0c6135f174c2d9d51b48b236c83964bd864f5cfd18979864 WatchSource:0}: Error finding container 57b012ebf793a59f0c6135f174c2d9d51b48b236c83964bd864f5cfd18979864: Status 404 returned error can't find the container with id 57b012ebf793a59f0c6135f174c2d9d51b48b236c83964bd864f5cfd18979864 Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.769584 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data" (OuterVolumeSpecName: "config-data") pod "7243da66-6008-464f-a5f0-423e342f34d0" (UID: "7243da66-6008-464f-a5f0-423e342f34d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.792072 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wglnl\" (UniqueName: \"kubernetes.io/projected/7243da66-6008-464f-a5f0-423e342f34d0-kube-api-access-wglnl\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.792208 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.792285 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.792363 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7243da66-6008-464f-a5f0-423e342f34d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.792426 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7243da66-6008-464f-a5f0-423e342f34d0-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:44 crc kubenswrapper[4837]: I1001 07:26:44.912842 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.613510 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerID="477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1" exitCode=2 Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.614500 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerID="e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af" exitCode=0 Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.613615 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerDied","Data":"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1"} Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.614592 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerDied","Data":"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af"} Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.618817 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerStarted","Data":"b1a25a5631787393515c8e82608962228bfb05a6623f191895b95414994acc1b"} Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.618856 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerStarted","Data":"57b012ebf793a59f0c6135f174c2d9d51b48b236c83964bd864f5cfd18979864"} Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.620866 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerStarted","Data":"a708b9492c517db3c4ac65ac6d8e06a0899c2e2382102fdd99be860d6cabafbe"} Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.620902 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b5db87b6d-sq97p" Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.664140 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.664119839 podStartE2EDuration="3.664119839s" podCreationTimestamp="2025-10-01 07:26:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:45.664105939 +0000 UTC m=+1262.505713414" watchObservedRunningTime="2025-10-01 07:26:45.664119839 +0000 UTC m=+1262.505727294" Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.694816 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.701662 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7b5db87b6d-sq97p"] Oct 01 07:26:45 crc kubenswrapper[4837]: I1001 07:26:45.829878 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7243da66-6008-464f-a5f0-423e342f34d0" path="/var/lib/kubelet/pods/7243da66-6008-464f-a5f0-423e342f34d0/volumes" Oct 01 07:26:46 crc kubenswrapper[4837]: I1001 07:26:46.636272 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerID="e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa" exitCode=0 Oct 01 07:26:46 crc kubenswrapper[4837]: I1001 07:26:46.636340 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerDied","Data":"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa"} Oct 01 07:26:46 crc kubenswrapper[4837]: I1001 07:26:46.640838 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerStarted","Data":"470c5f7992ab6223b3d21e3422fa5783032efa7c901479e728e382c4209680ac"} Oct 01 07:26:46 crc kubenswrapper[4837]: I1001 07:26:46.694182 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.694164246 podStartE2EDuration="3.694164246s" podCreationTimestamp="2025-10-01 07:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:46.682165769 +0000 UTC m=+1263.523773224" watchObservedRunningTime="2025-10-01 07:26:46.694164246 +0000 UTC m=+1263.535771701" Oct 01 07:26:50 crc kubenswrapper[4837]: I1001 07:26:50.151165 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.685851 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c192-account-create-52hlt"] Oct 01 07:26:52 crc kubenswrapper[4837]: E1001 07:26:52.687112 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.687137 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api" Oct 01 07:26:52 crc kubenswrapper[4837]: E1001 07:26:52.687166 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api-log" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.687179 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api-log" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.687465 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.687513 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7243da66-6008-464f-a5f0-423e342f34d0" containerName="barbican-api-log" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.688529 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.693305 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.702777 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c192-account-create-52hlt"] Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.753836 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfhh\" (UniqueName: \"kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh\") pod \"nova-api-c192-account-create-52hlt\" (UID: \"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654\") " pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.855242 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfhh\" (UniqueName: \"kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh\") pod \"nova-api-c192-account-create-52hlt\" (UID: \"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654\") " pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.870198 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-fbd8-account-create-5wlnj"] Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.871986 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.874763 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.887195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfhh\" (UniqueName: \"kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh\") pod \"nova-api-c192-account-create-52hlt\" (UID: \"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654\") " pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.898823 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fbd8-account-create-5wlnj"] Oct 01 07:26:52 crc kubenswrapper[4837]: I1001 07:26:52.957168 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7wpb\" (UniqueName: \"kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb\") pod \"nova-cell0-fbd8-account-create-5wlnj\" (UID: \"abc5c46e-5e15-4c96-b1d8-7867068c488d\") " pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.019631 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.019887 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.030990 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.058830 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7wpb\" (UniqueName: \"kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb\") pod \"nova-cell0-fbd8-account-create-5wlnj\" (UID: \"abc5c46e-5e15-4c96-b1d8-7867068c488d\") " pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.071337 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.079533 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.079586 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.080598 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1e55-account-create-7588j"] Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.085190 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.085889 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.089281 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.105633 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7wpb\" (UniqueName: \"kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb\") pod \"nova-cell0-fbd8-account-create-5wlnj\" (UID: \"abc5c46e-5e15-4c96-b1d8-7867068c488d\") " pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.127375 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1e55-account-create-7588j"] Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.160305 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4q7p\" (UniqueName: \"kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p\") pod \"nova-cell1-1e55-account-create-7588j\" (UID: \"d20e80bb-46a9-4b04-b2d3-132683af9fb3\") " pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.260161 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.261901 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4q7p\" (UniqueName: \"kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p\") pod \"nova-cell1-1e55-account-create-7588j\" (UID: \"d20e80bb-46a9-4b04-b2d3-132683af9fb3\") " pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.283601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4q7p\" (UniqueName: \"kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p\") pod \"nova-cell1-1e55-account-create-7588j\" (UID: \"d20e80bb-46a9-4b04-b2d3-132683af9fb3\") " pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.508707 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.526199 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c192-account-create-52hlt"] Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.741421 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fbd8-account-create-5wlnj"] Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.743606 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c192-account-create-52hlt" event={"ID":"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654","Type":"ContainerStarted","Data":"d87ec853871ad0d44a4f50125b836c1f87633551c852c805248ea18a8a04fa83"} Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.743643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c192-account-create-52hlt" event={"ID":"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654","Type":"ContainerStarted","Data":"30a6f4d8a149104b11432952f4aa5d6cb2bd2b715b8437b97738b82d16374cb2"} Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.744149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.745267 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: W1001 07:26:53.748421 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabc5c46e_5e15_4c96_b1d8_7867068c488d.slice/crio-ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99 WatchSource:0}: Error finding container ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99: Status 404 returned error can't find the container with id ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99 Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.770977 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c192-account-create-52hlt" podStartSLOduration=1.770956466 podStartE2EDuration="1.770956466s" podCreationTimestamp="2025-10-01 07:26:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:26:53.764346323 +0000 UTC m=+1270.605953778" watchObservedRunningTime="2025-10-01 07:26:53.770956466 +0000 UTC m=+1270.612563941" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.959679 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1e55-account-create-7588j"] Oct 01 07:26:53 crc kubenswrapper[4837]: W1001 07:26:53.969350 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd20e80bb_46a9_4b04_b2d3_132683af9fb3.slice/crio-750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f WatchSource:0}: Error finding container 750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f: Status 404 returned error can't find the container with id 750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.988800 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:53 crc kubenswrapper[4837]: I1001 07:26:53.988842 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.038079 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.038414 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:54 crc kubenswrapper[4837]: E1001 07:26:54.494374 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd20e80bb_46a9_4b04_b2d3_132683af9fb3.slice/crio-9dfecee7787a639e916f3c2e6d2900c5192cf19d000491e365456ae7ad50656d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd20e80bb_46a9_4b04_b2d3_132683af9fb3.slice/crio-conmon-9dfecee7787a639e916f3c2e6d2900c5192cf19d000491e365456ae7ad50656d.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.757576 4837 generic.go:334] "Generic (PLEG): container finished" podID="d20e80bb-46a9-4b04-b2d3-132683af9fb3" containerID="9dfecee7787a639e916f3c2e6d2900c5192cf19d000491e365456ae7ad50656d" exitCode=0 Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.757796 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1e55-account-create-7588j" event={"ID":"d20e80bb-46a9-4b04-b2d3-132683af9fb3","Type":"ContainerDied","Data":"9dfecee7787a639e916f3c2e6d2900c5192cf19d000491e365456ae7ad50656d"} Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.760128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1e55-account-create-7588j" event={"ID":"d20e80bb-46a9-4b04-b2d3-132683af9fb3","Type":"ContainerStarted","Data":"750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f"} Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.762901 4837 generic.go:334] "Generic (PLEG): container finished" podID="99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" containerID="d87ec853871ad0d44a4f50125b836c1f87633551c852c805248ea18a8a04fa83" exitCode=0 Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.763166 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c192-account-create-52hlt" event={"ID":"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654","Type":"ContainerDied","Data":"d87ec853871ad0d44a4f50125b836c1f87633551c852c805248ea18a8a04fa83"} Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.765539 4837 generic.go:334] "Generic (PLEG): container finished" podID="abc5c46e-5e15-4c96-b1d8-7867068c488d" containerID="3552d4b1a01b53a4ee56dd82ae9cca651bd961e33681b9c72d8f9260300ac389" exitCode=0 Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.765751 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" event={"ID":"abc5c46e-5e15-4c96-b1d8-7867068c488d","Type":"ContainerDied","Data":"3552d4b1a01b53a4ee56dd82ae9cca651bd961e33681b9c72d8f9260300ac389"} Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.765923 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" event={"ID":"abc5c46e-5e15-4c96-b1d8-7867068c488d","Type":"ContainerStarted","Data":"ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99"} Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.766229 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:54 crc kubenswrapper[4837]: I1001 07:26:54.766307 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:55 crc kubenswrapper[4837]: I1001 07:26:55.654292 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 07:26:55 crc kubenswrapper[4837]: I1001 07:26:55.654369 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.125866 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.219405 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7wpb\" (UniqueName: \"kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb\") pod \"abc5c46e-5e15-4c96-b1d8-7867068c488d\" (UID: \"abc5c46e-5e15-4c96-b1d8-7867068c488d\") " Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.226276 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb" (OuterVolumeSpecName: "kube-api-access-s7wpb") pod "abc5c46e-5e15-4c96-b1d8-7867068c488d" (UID: "abc5c46e-5e15-4c96-b1d8-7867068c488d"). InnerVolumeSpecName "kube-api-access-s7wpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.275582 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.281830 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.322429 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7wpb\" (UniqueName: \"kubernetes.io/projected/abc5c46e-5e15-4c96-b1d8-7867068c488d-kube-api-access-s7wpb\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.423825 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4q7p\" (UniqueName: \"kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p\") pod \"d20e80bb-46a9-4b04-b2d3-132683af9fb3\" (UID: \"d20e80bb-46a9-4b04-b2d3-132683af9fb3\") " Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.423964 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdfhh\" (UniqueName: \"kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh\") pod \"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654\" (UID: \"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654\") " Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.428741 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p" (OuterVolumeSpecName: "kube-api-access-w4q7p") pod "d20e80bb-46a9-4b04-b2d3-132683af9fb3" (UID: "d20e80bb-46a9-4b04-b2d3-132683af9fb3"). InnerVolumeSpecName "kube-api-access-w4q7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.429326 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh" (OuterVolumeSpecName: "kube-api-access-sdfhh") pod "99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" (UID: "99acf906-a3d0-4aeb-9fc8-3c9f9ba28654"). InnerVolumeSpecName "kube-api-access-sdfhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.526180 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4q7p\" (UniqueName: \"kubernetes.io/projected/d20e80bb-46a9-4b04-b2d3-132683af9fb3-kube-api-access-w4q7p\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.526211 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdfhh\" (UniqueName: \"kubernetes.io/projected/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654-kube-api-access-sdfhh\") on node \"crc\" DevicePath \"\"" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.638056 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.671641 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.792500 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c192-account-create-52hlt" event={"ID":"99acf906-a3d0-4aeb-9fc8-3c9f9ba28654","Type":"ContainerDied","Data":"30a6f4d8a149104b11432952f4aa5d6cb2bd2b715b8437b97738b82d16374cb2"} Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.792540 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30a6f4d8a149104b11432952f4aa5d6cb2bd2b715b8437b97738b82d16374cb2" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.792588 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c192-account-create-52hlt" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.798144 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" event={"ID":"abc5c46e-5e15-4c96-b1d8-7867068c488d","Type":"ContainerDied","Data":"ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99"} Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.798195 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fbd8-account-create-5wlnj" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.798205 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed3ca637a11209f62f802e3e9c41d7228c4388446a3b509ab6740ab3ece75f99" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.803561 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1e55-account-create-7588j" Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.803758 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1e55-account-create-7588j" event={"ID":"d20e80bb-46a9-4b04-b2d3-132683af9fb3","Type":"ContainerDied","Data":"750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f"} Oct 01 07:26:56 crc kubenswrapper[4837]: I1001 07:26:56.803800 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="750b1c02000b1a3a45cc69be725f743977d51dfe72b2992d65269b3128ece32f" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.130503 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vd4tb"] Oct 01 07:26:58 crc kubenswrapper[4837]: E1001 07:26:58.131117 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20e80bb-46a9-4b04-b2d3-132683af9fb3" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131129 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20e80bb-46a9-4b04-b2d3-132683af9fb3" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: E1001 07:26:58.131146 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc5c46e-5e15-4c96-b1d8-7867068c488d" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131154 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc5c46e-5e15-4c96-b1d8-7867068c488d" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: E1001 07:26:58.131169 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131175 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131321 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20e80bb-46a9-4b04-b2d3-132683af9fb3" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131339 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131348 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc5c46e-5e15-4c96-b1d8-7867068c488d" containerName="mariadb-account-create" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.131941 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.133772 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vhsq2" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.134069 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.135474 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.151537 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vd4tb"] Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.265273 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hl5m\" (UniqueName: \"kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.265401 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.265436 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.265453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.366857 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.366945 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.366983 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.367091 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hl5m\" (UniqueName: \"kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.374223 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.374372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.375068 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.388433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hl5m\" (UniqueName: \"kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m\") pod \"nova-cell0-conductor-db-sync-vd4tb\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.450935 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:26:58 crc kubenswrapper[4837]: W1001 07:26:58.883454 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9681804a_3e4e_44a1_b1fc_8280c44806bd.slice/crio-b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b WatchSource:0}: Error finding container b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b: Status 404 returned error can't find the container with id b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b Oct 01 07:26:58 crc kubenswrapper[4837]: I1001 07:26:58.897036 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vd4tb"] Oct 01 07:26:59 crc kubenswrapper[4837]: I1001 07:26:59.851165 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" event={"ID":"9681804a-3e4e-44a1-b1fc-8280c44806bd","Type":"ContainerStarted","Data":"b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b"} Oct 01 07:27:06 crc kubenswrapper[4837]: I1001 07:27:06.932417 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" event={"ID":"9681804a-3e4e-44a1-b1fc-8280c44806bd","Type":"ContainerStarted","Data":"af16c63816c1328370401333ff150789d7df93335939e3bb0c6ec880848f98b8"} Oct 01 07:27:06 crc kubenswrapper[4837]: I1001 07:27:06.954603 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" podStartSLOduration=1.881280324 podStartE2EDuration="8.954580229s" podCreationTimestamp="2025-10-01 07:26:58 +0000 UTC" firstStartedPulling="2025-10-01 07:26:58.887129995 +0000 UTC m=+1275.728737450" lastFinishedPulling="2025-10-01 07:27:05.96042991 +0000 UTC m=+1282.802037355" observedRunningTime="2025-10-01 07:27:06.953753859 +0000 UTC m=+1283.795361354" watchObservedRunningTime="2025-10-01 07:27:06.954580229 +0000 UTC m=+1283.796187684" Oct 01 07:27:09 crc kubenswrapper[4837]: I1001 07:27:09.879008 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 01 07:27:14 crc kubenswrapper[4837]: I1001 07:27:14.967272 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.034013 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerID="b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654" exitCode=137 Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.034285 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerDied","Data":"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654"} Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.034316 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2e800fd-f1e2-46e2-a806-d17c54adb0a2","Type":"ContainerDied","Data":"731d326c4923f1a7071c0c258cbba267ab57d637b9f99801c2d3e608a252b61c"} Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.034339 4837 scope.go:117] "RemoveContainer" containerID="b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.034488 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.054844 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.054931 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.054992 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055021 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9cbb\" (UniqueName: \"kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055056 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055185 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055209 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle\") pod \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\" (UID: \"d2e800fd-f1e2-46e2-a806-d17c54adb0a2\") " Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055892 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.055939 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.057261 4837 scope.go:117] "RemoveContainer" containerID="477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.060602 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts" (OuterVolumeSpecName: "scripts") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.061597 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb" (OuterVolumeSpecName: "kube-api-access-z9cbb") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "kube-api-access-z9cbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.100461 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.154402 4837 scope.go:117] "RemoveContainer" containerID="e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.156770 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.156798 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.156807 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.156815 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.156825 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9cbb\" (UniqueName: \"kubernetes.io/projected/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-kube-api-access-z9cbb\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.158614 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.178820 4837 scope.go:117] "RemoveContainer" containerID="e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.180231 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data" (OuterVolumeSpecName: "config-data") pod "d2e800fd-f1e2-46e2-a806-d17c54adb0a2" (UID: "d2e800fd-f1e2-46e2-a806-d17c54adb0a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.201960 4837 scope.go:117] "RemoveContainer" containerID="b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.202268 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654\": container with ID starting with b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654 not found: ID does not exist" containerID="b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202296 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654"} err="failed to get container status \"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654\": rpc error: code = NotFound desc = could not find container \"b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654\": container with ID starting with b1ce7e5f4387a95b0f15f02dd27ddda2503121bc91386d81121f3cc01f9c1654 not found: ID does not exist" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202320 4837 scope.go:117] "RemoveContainer" containerID="477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.202516 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1\": container with ID starting with 477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1 not found: ID does not exist" containerID="477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202538 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1"} err="failed to get container status \"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1\": rpc error: code = NotFound desc = could not find container \"477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1\": container with ID starting with 477a7afe65ac374ca274d7d1d124d87d2862028aa4dc6281ed358e65cfd340b1 not found: ID does not exist" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202552 4837 scope.go:117] "RemoveContainer" containerID="e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.202754 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af\": container with ID starting with e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af not found: ID does not exist" containerID="e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202774 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af"} err="failed to get container status \"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af\": rpc error: code = NotFound desc = could not find container \"e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af\": container with ID starting with e3aa587fa509c5a4bc0e794baaea48f169128d14ec3b44bb424718fbdd9e47af not found: ID does not exist" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202786 4837 scope.go:117] "RemoveContainer" containerID="e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.202945 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa\": container with ID starting with e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa not found: ID does not exist" containerID="e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.202966 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa"} err="failed to get container status \"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa\": rpc error: code = NotFound desc = could not find container \"e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa\": container with ID starting with e7f20a6a35884f972fd03d9ed6ec9108a39a9f8f0f5326c18b45ad83c80aa6aa not found: ID does not exist" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.258442 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.258480 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e800fd-f1e2-46e2-a806-d17c54adb0a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.368432 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.380704 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.404437 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.404781 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-central-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.404794 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-central-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.404807 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-notification-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.404814 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-notification-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.404830 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="sg-core" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.404836 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="sg-core" Oct 01 07:27:15 crc kubenswrapper[4837]: E1001 07:27:15.404857 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="proxy-httpd" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.404862 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="proxy-httpd" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.405028 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-central-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.405049 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="sg-core" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.405059 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="ceilometer-notification-agent" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.405071 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" containerName="proxy-httpd" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.406521 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.410282 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.412432 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.416047 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.568727 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.568794 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.568832 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.568967 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.569315 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.569622 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.569757 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp962\" (UniqueName: \"kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672220 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672564 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672592 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672625 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672743 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672773 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.672796 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp962\" (UniqueName: \"kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.673337 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.674008 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.677170 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.677278 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.677678 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.690266 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.709072 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp962\" (UniqueName: \"kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962\") pod \"ceilometer-0\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.745330 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:15 crc kubenswrapper[4837]: I1001 07:27:15.835207 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e800fd-f1e2-46e2-a806-d17c54adb0a2" path="/var/lib/kubelet/pods/d2e800fd-f1e2-46e2-a806-d17c54adb0a2/volumes" Oct 01 07:27:16 crc kubenswrapper[4837]: I1001 07:27:16.246264 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:17 crc kubenswrapper[4837]: I1001 07:27:17.061641 4837 generic.go:334] "Generic (PLEG): container finished" podID="9681804a-3e4e-44a1-b1fc-8280c44806bd" containerID="af16c63816c1328370401333ff150789d7df93335939e3bb0c6ec880848f98b8" exitCode=0 Oct 01 07:27:17 crc kubenswrapper[4837]: I1001 07:27:17.061934 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" event={"ID":"9681804a-3e4e-44a1-b1fc-8280c44806bd","Type":"ContainerDied","Data":"af16c63816c1328370401333ff150789d7df93335939e3bb0c6ec880848f98b8"} Oct 01 07:27:17 crc kubenswrapper[4837]: I1001 07:27:17.067368 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerStarted","Data":"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e"} Oct 01 07:27:17 crc kubenswrapper[4837]: I1001 07:27:17.067410 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerStarted","Data":"3acd5b7fa9c5da9e99678f4a8ec86451cdd10b893899312f908c083b4c216abf"} Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.103005 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerStarted","Data":"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6"} Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.463708 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.627442 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts\") pod \"9681804a-3e4e-44a1-b1fc-8280c44806bd\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.627623 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle\") pod \"9681804a-3e4e-44a1-b1fc-8280c44806bd\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.627659 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data\") pod \"9681804a-3e4e-44a1-b1fc-8280c44806bd\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.627772 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hl5m\" (UniqueName: \"kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m\") pod \"9681804a-3e4e-44a1-b1fc-8280c44806bd\" (UID: \"9681804a-3e4e-44a1-b1fc-8280c44806bd\") " Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.631835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts" (OuterVolumeSpecName: "scripts") pod "9681804a-3e4e-44a1-b1fc-8280c44806bd" (UID: "9681804a-3e4e-44a1-b1fc-8280c44806bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.632814 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m" (OuterVolumeSpecName: "kube-api-access-5hl5m") pod "9681804a-3e4e-44a1-b1fc-8280c44806bd" (UID: "9681804a-3e4e-44a1-b1fc-8280c44806bd"). InnerVolumeSpecName "kube-api-access-5hl5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.653839 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data" (OuterVolumeSpecName: "config-data") pod "9681804a-3e4e-44a1-b1fc-8280c44806bd" (UID: "9681804a-3e4e-44a1-b1fc-8280c44806bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.670748 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9681804a-3e4e-44a1-b1fc-8280c44806bd" (UID: "9681804a-3e4e-44a1-b1fc-8280c44806bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.730006 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.730300 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.730321 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hl5m\" (UniqueName: \"kubernetes.io/projected/9681804a-3e4e-44a1-b1fc-8280c44806bd-kube-api-access-5hl5m\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:18 crc kubenswrapper[4837]: I1001 07:27:18.730339 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9681804a-3e4e-44a1-b1fc-8280c44806bd-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.117820 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerStarted","Data":"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705"} Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.119653 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" event={"ID":"9681804a-3e4e-44a1-b1fc-8280c44806bd","Type":"ContainerDied","Data":"b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b"} Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.119754 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7832d2454af3c63f00d5e36423bbf04ea3a0e541c7a9c72c50edf03d17b742b" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.119815 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vd4tb" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.223993 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:27:19 crc kubenswrapper[4837]: E1001 07:27:19.224440 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9681804a-3e4e-44a1-b1fc-8280c44806bd" containerName="nova-cell0-conductor-db-sync" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.224467 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9681804a-3e4e-44a1-b1fc-8280c44806bd" containerName="nova-cell0-conductor-db-sync" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.224750 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9681804a-3e4e-44a1-b1fc-8280c44806bd" containerName="nova-cell0-conductor-db-sync" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.225424 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.228619 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vhsq2" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.228970 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.234550 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.352430 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.352481 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.352887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htgkt\" (UniqueName: \"kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.454584 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.454640 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.454779 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htgkt\" (UniqueName: \"kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.458085 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.458804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.478375 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htgkt\" (UniqueName: \"kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt\") pod \"nova-cell0-conductor-0\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:19 crc kubenswrapper[4837]: I1001 07:27:19.555833 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:20 crc kubenswrapper[4837]: I1001 07:27:20.033332 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:27:20 crc kubenswrapper[4837]: I1001 07:27:20.132557 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c09c611-b859-4650-b4b0-96738b65fc45","Type":"ContainerStarted","Data":"acfb24c0dc96b2ced53b34659ef44aeb135e57d07544dc84f4597c41be3f48d9"} Oct 01 07:27:20 crc kubenswrapper[4837]: I1001 07:27:20.136006 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerStarted","Data":"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33"} Oct 01 07:27:20 crc kubenswrapper[4837]: I1001 07:27:20.136919 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:27:20 crc kubenswrapper[4837]: I1001 07:27:20.162237 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.009149566 podStartE2EDuration="5.162220875s" podCreationTimestamp="2025-10-01 07:27:15 +0000 UTC" firstStartedPulling="2025-10-01 07:27:16.252569519 +0000 UTC m=+1293.094177004" lastFinishedPulling="2025-10-01 07:27:19.405640858 +0000 UTC m=+1296.247248313" observedRunningTime="2025-10-01 07:27:20.161851167 +0000 UTC m=+1297.003458632" watchObservedRunningTime="2025-10-01 07:27:20.162220875 +0000 UTC m=+1297.003828330" Oct 01 07:27:21 crc kubenswrapper[4837]: I1001 07:27:21.176337 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c09c611-b859-4650-b4b0-96738b65fc45","Type":"ContainerStarted","Data":"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8"} Oct 01 07:27:21 crc kubenswrapper[4837]: I1001 07:27:21.184119 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:21 crc kubenswrapper[4837]: I1001 07:27:21.214842 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.214821655 podStartE2EDuration="2.214821655s" podCreationTimestamp="2025-10-01 07:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:21.206317334 +0000 UTC m=+1298.047924799" watchObservedRunningTime="2025-10-01 07:27:21.214821655 +0000 UTC m=+1298.056429120" Oct 01 07:27:23 crc kubenswrapper[4837]: I1001 07:27:23.080319 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:27:23 crc kubenswrapper[4837]: I1001 07:27:23.080377 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:27:29 crc kubenswrapper[4837]: I1001 07:27:29.600299 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.142866 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-l5h6t"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.145207 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.148919 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.149186 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.153618 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-l5h6t"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.269979 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.270111 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.270207 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.270269 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqvws\" (UniqueName: \"kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.318596 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.321655 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.327854 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.334261 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.372191 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.372261 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.372292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.372322 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqvws\" (UniqueName: \"kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.379132 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.383899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.384078 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.389864 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqvws\" (UniqueName: \"kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws\") pod \"nova-cell0-cell-mapping-l5h6t\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.425829 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.429154 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.431194 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.439384 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.474190 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8655\" (UniqueName: \"kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.474318 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.474374 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.495083 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.496249 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.499617 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.512046 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.516026 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.569839 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.571366 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.577075 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.577780 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578001 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578040 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578081 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8655\" (UniqueName: \"kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqv55\" (UniqueName: \"kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578508 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578524 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmfn4\" (UniqueName: \"kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.578562 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.582186 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.584889 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.590405 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.619117 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8655\" (UniqueName: \"kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655\") pod \"nova-cell1-novncproxy-0\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.638454 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.642970 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.644443 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679729 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jhzw\" (UniqueName: \"kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679814 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679844 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679876 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679914 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679954 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679972 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqv55\" (UniqueName: \"kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.679999 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.680020 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.680050 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmfn4\" (UniqueName: \"kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.680661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.687460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.688413 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.690560 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.697026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.700494 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.700512 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmfn4\" (UniqueName: \"kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4\") pod \"nova-api-0\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.701785 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqv55\" (UniqueName: \"kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55\") pod \"nova-scheduler-0\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.777858 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783063 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783350 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783368 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvsp\" (UniqueName: \"kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783395 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783452 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783490 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jhzw\" (UniqueName: \"kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783530 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.783578 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.784009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.788187 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.788239 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.804328 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jhzw\" (UniqueName: \"kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw\") pod \"nova-metadata-0\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " pod="openstack/nova-metadata-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.836546 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885701 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885814 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvsp\" (UniqueName: \"kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885884 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885950 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.885996 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.886487 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.886918 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.889292 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.889575 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.889778 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:30 crc kubenswrapper[4837]: I1001 07:27:30.902334 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvsp\" (UniqueName: \"kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp\") pod \"dnsmasq-dns-6f5d75d8c-7z2cd\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.014766 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.035413 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.156406 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-l5h6t"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.166466 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.303221 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b82e1abf-77fe-498c-8a09-2e509a551607","Type":"ContainerStarted","Data":"11dd6ec3ed9f649bf0d5836beffb32c9ceec3b315ccc784b55881dc75e8bac14"} Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.305660 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-l5h6t" event={"ID":"a1ca37a0-8611-42e6-b365-7ed2aa8dc811","Type":"ContainerStarted","Data":"685b2ba643278077e23abd34c7ad6e6fff96f0cf27e068edbae666856d6e10f9"} Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.327454 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lrdgf"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.328649 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.333500 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.333800 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.347659 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.361754 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lrdgf"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.395512 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.395808 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.395859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f76mg\" (UniqueName: \"kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.395919 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.454335 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.497889 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.497955 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f76mg\" (UniqueName: \"kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.497990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.498059 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.502989 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.504394 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.505621 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.520934 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f76mg\" (UniqueName: \"kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg\") pod \"nova-cell1-conductor-db-sync-lrdgf\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.658501 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.662734 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:31 crc kubenswrapper[4837]: I1001 07:27:31.733796 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.104832 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lrdgf"] Oct 01 07:27:32 crc kubenswrapper[4837]: W1001 07:27:32.113721 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef375f9_54f9_4c27_88c8_8a8345d4a5f8.slice/crio-e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17 WatchSource:0}: Error finding container e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17: Status 404 returned error can't find the container with id e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17 Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.316861 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerStarted","Data":"ee7e9e2c03e286a5d4f0de649b7b34e4430f77fe09d406ef3d41cd89149c29be"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.321745 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"387b138c-ac22-411d-9692-41f7146df692","Type":"ContainerStarted","Data":"d58bc394415238a254ccf79511faf5c6019ad590722b2273c23e8ee6ab28e817"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.323212 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerStarted","Data":"d72640bab6def49bbe8c0acb5688295c9f2fd5b4f6eb4ab3aaa91e6762572ed5"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.326189 4837 generic.go:334] "Generic (PLEG): container finished" podID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerID="7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57" exitCode=0 Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.326248 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" event={"ID":"1c7351cc-3937-4008-a199-5fe7ccaa6276","Type":"ContainerDied","Data":"7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.326266 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" event={"ID":"1c7351cc-3937-4008-a199-5fe7ccaa6276","Type":"ContainerStarted","Data":"1ff26642dab94cc1b8980302658e5c2f6625d83998967aaeb0abfc38d72f6ce5"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.339462 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-l5h6t" event={"ID":"a1ca37a0-8611-42e6-b365-7ed2aa8dc811","Type":"ContainerStarted","Data":"8b0122a9f0b165396dc68834bcd670b9d796dd63c3cd41cd5cc6e187caf12ecf"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.341792 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" event={"ID":"aef375f9-54f9-4c27-88c8-8a8345d4a5f8","Type":"ContainerStarted","Data":"e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17"} Oct 01 07:27:32 crc kubenswrapper[4837]: I1001 07:27:32.379834 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-l5h6t" podStartSLOduration=2.379809908 podStartE2EDuration="2.379809908s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:32.362684223 +0000 UTC m=+1309.204291678" watchObservedRunningTime="2025-10-01 07:27:32.379809908 +0000 UTC m=+1309.221417373" Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.355803 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" event={"ID":"aef375f9-54f9-4c27-88c8-8a8345d4a5f8","Type":"ContainerStarted","Data":"8b0403847fa1e913e31eb2c874d46883eb7f9de350b911b8a7adf2ee314d096a"} Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.362110 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" event={"ID":"1c7351cc-3937-4008-a199-5fe7ccaa6276","Type":"ContainerStarted","Data":"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08"} Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.362141 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.379340 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" podStartSLOduration=2.379319189 podStartE2EDuration="2.379319189s" podCreationTimestamp="2025-10-01 07:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:33.368721266 +0000 UTC m=+1310.210328721" watchObservedRunningTime="2025-10-01 07:27:33.379319189 +0000 UTC m=+1310.220926644" Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.395494 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" podStartSLOduration=3.395476411 podStartE2EDuration="3.395476411s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:33.387643426 +0000 UTC m=+1310.229250901" watchObservedRunningTime="2025-10-01 07:27:33.395476411 +0000 UTC m=+1310.237083866" Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.904375 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:27:33 crc kubenswrapper[4837]: I1001 07:27:33.910532 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.387794 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerStarted","Data":"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.388533 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerStarted","Data":"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.391059 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerStarted","Data":"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.391096 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerStarted","Data":"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.391278 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-log" containerID="cri-o://c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" gracePeriod=30 Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.391329 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-metadata" containerID="cri-o://8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" gracePeriod=30 Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.396823 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b82e1abf-77fe-498c-8a09-2e509a551607","Type":"ContainerStarted","Data":"7e1bae08709298165b16e91922f8d331270e19bce0e857867200a431bc764409"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.397167 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b82e1abf-77fe-498c-8a09-2e509a551607" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7e1bae08709298165b16e91922f8d331270e19bce0e857867200a431bc764409" gracePeriod=30 Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.403575 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"387b138c-ac22-411d-9692-41f7146df692","Type":"ContainerStarted","Data":"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a"} Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.413026 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.492373683 podStartE2EDuration="5.413002072s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="2025-10-01 07:27:31.353427139 +0000 UTC m=+1308.195034584" lastFinishedPulling="2025-10-01 07:27:34.274055508 +0000 UTC m=+1311.115662973" observedRunningTime="2025-10-01 07:27:35.411006142 +0000 UTC m=+1312.252613597" watchObservedRunningTime="2025-10-01 07:27:35.413002072 +0000 UTC m=+1312.254609567" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.450680 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.371283287 podStartE2EDuration="5.450664567s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="2025-10-01 07:27:31.19477708 +0000 UTC m=+1308.036384535" lastFinishedPulling="2025-10-01 07:27:34.27415836 +0000 UTC m=+1311.115765815" observedRunningTime="2025-10-01 07:27:35.427128573 +0000 UTC m=+1312.268736028" watchObservedRunningTime="2025-10-01 07:27:35.450664567 +0000 UTC m=+1312.292272022" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.453831 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.644541022 podStartE2EDuration="5.453824186s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="2025-10-01 07:27:31.46300715 +0000 UTC m=+1308.304614595" lastFinishedPulling="2025-10-01 07:27:34.272290304 +0000 UTC m=+1311.113897759" observedRunningTime="2025-10-01 07:27:35.438779022 +0000 UTC m=+1312.280386477" watchObservedRunningTime="2025-10-01 07:27:35.453824186 +0000 UTC m=+1312.295431641" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.466865 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.874538164 podStartE2EDuration="5.466845299s" podCreationTimestamp="2025-10-01 07:27:30 +0000 UTC" firstStartedPulling="2025-10-01 07:27:31.68002159 +0000 UTC m=+1308.521629045" lastFinishedPulling="2025-10-01 07:27:34.272328725 +0000 UTC m=+1311.113936180" observedRunningTime="2025-10-01 07:27:35.464844529 +0000 UTC m=+1312.306451984" watchObservedRunningTime="2025-10-01 07:27:35.466845299 +0000 UTC m=+1312.308452754" Oct 01 07:27:35 crc kubenswrapper[4837]: E1001 07:27:35.551121 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode16d452e_cf98_48c8_bde7_f003276bfa37.slice/crio-conmon-c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode16d452e_cf98_48c8_bde7_f003276bfa37.slice/crio-c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.638879 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.837624 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.935870 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.983314 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs\") pod \"e16d452e-cf98-48c8-bde7-f003276bfa37\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.983462 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jhzw\" (UniqueName: \"kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw\") pod \"e16d452e-cf98-48c8-bde7-f003276bfa37\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.983597 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data\") pod \"e16d452e-cf98-48c8-bde7-f003276bfa37\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.983645 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle\") pod \"e16d452e-cf98-48c8-bde7-f003276bfa37\" (UID: \"e16d452e-cf98-48c8-bde7-f003276bfa37\") " Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.983860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs" (OuterVolumeSpecName: "logs") pod "e16d452e-cf98-48c8-bde7-f003276bfa37" (UID: "e16d452e-cf98-48c8-bde7-f003276bfa37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.984612 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e16d452e-cf98-48c8-bde7-f003276bfa37-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:35 crc kubenswrapper[4837]: I1001 07:27:35.989595 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw" (OuterVolumeSpecName: "kube-api-access-9jhzw") pod "e16d452e-cf98-48c8-bde7-f003276bfa37" (UID: "e16d452e-cf98-48c8-bde7-f003276bfa37"). InnerVolumeSpecName "kube-api-access-9jhzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.018178 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e16d452e-cf98-48c8-bde7-f003276bfa37" (UID: "e16d452e-cf98-48c8-bde7-f003276bfa37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.022954 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data" (OuterVolumeSpecName: "config-data") pod "e16d452e-cf98-48c8-bde7-f003276bfa37" (UID: "e16d452e-cf98-48c8-bde7-f003276bfa37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.086199 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jhzw\" (UniqueName: \"kubernetes.io/projected/e16d452e-cf98-48c8-bde7-f003276bfa37-kube-api-access-9jhzw\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.086228 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.086238 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d452e-cf98-48c8-bde7-f003276bfa37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.420752 4837 generic.go:334] "Generic (PLEG): container finished" podID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerID="8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" exitCode=0 Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.420798 4837 generic.go:334] "Generic (PLEG): container finished" podID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerID="c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" exitCode=143 Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.421073 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerDied","Data":"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d"} Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.421134 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerDied","Data":"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac"} Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.421148 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e16d452e-cf98-48c8-bde7-f003276bfa37","Type":"ContainerDied","Data":"ee7e9e2c03e286a5d4f0de649b7b34e4430f77fe09d406ef3d41cd89149c29be"} Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.421167 4837 scope.go:117] "RemoveContainer" containerID="8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.421781 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.475259 4837 scope.go:117] "RemoveContainer" containerID="c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.488760 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.512118 4837 scope.go:117] "RemoveContainer" containerID="8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.512306 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:36 crc kubenswrapper[4837]: E1001 07:27:36.513025 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d\": container with ID starting with 8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d not found: ID does not exist" containerID="8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.513071 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d"} err="failed to get container status \"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d\": rpc error: code = NotFound desc = could not find container \"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d\": container with ID starting with 8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d not found: ID does not exist" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.513106 4837 scope.go:117] "RemoveContainer" containerID="c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" Oct 01 07:27:36 crc kubenswrapper[4837]: E1001 07:27:36.513622 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac\": container with ID starting with c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac not found: ID does not exist" containerID="c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.513664 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac"} err="failed to get container status \"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac\": rpc error: code = NotFound desc = could not find container \"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac\": container with ID starting with c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac not found: ID does not exist" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.513721 4837 scope.go:117] "RemoveContainer" containerID="8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.514114 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d"} err="failed to get container status \"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d\": rpc error: code = NotFound desc = could not find container \"8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d\": container with ID starting with 8d339931ea7ee9d896050637e6ee140dbe2e7cee82ecb6fd0258f2d1fe097a7d not found: ID does not exist" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.514151 4837 scope.go:117] "RemoveContainer" containerID="c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.520478 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac"} err="failed to get container status \"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac\": rpc error: code = NotFound desc = could not find container \"c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac\": container with ID starting with c7a15344ded1991499976bc1da50b14824ca81af93c36559845cd968bf28b6ac not found: ID does not exist" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.528376 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:36 crc kubenswrapper[4837]: E1001 07:27:36.529074 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-log" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.529109 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-log" Oct 01 07:27:36 crc kubenswrapper[4837]: E1001 07:27:36.529154 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-metadata" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.529165 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-metadata" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.529427 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-log" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.529464 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" containerName="nova-metadata-metadata" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.538363 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.546242 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.546507 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.559717 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.594112 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.594182 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.594215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.594260 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjm89\" (UniqueName: \"kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.594282 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.697026 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.697114 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.697183 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjm89\" (UniqueName: \"kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.697255 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.697806 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.698017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.711655 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.711888 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.712298 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.735131 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjm89\" (UniqueName: \"kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89\") pod \"nova-metadata-0\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " pod="openstack/nova-metadata-0" Oct 01 07:27:36 crc kubenswrapper[4837]: I1001 07:27:36.862209 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:37 crc kubenswrapper[4837]: I1001 07:27:37.356227 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:37 crc kubenswrapper[4837]: W1001 07:27:37.365287 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1258cc9c_10cb_4807_91de_e38be12f7533.slice/crio-00b642ab103d15b5096546a52d72540ad0852cb975bf0e4e8876f1050e8240bc WatchSource:0}: Error finding container 00b642ab103d15b5096546a52d72540ad0852cb975bf0e4e8876f1050e8240bc: Status 404 returned error can't find the container with id 00b642ab103d15b5096546a52d72540ad0852cb975bf0e4e8876f1050e8240bc Oct 01 07:27:37 crc kubenswrapper[4837]: I1001 07:27:37.432906 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerStarted","Data":"00b642ab103d15b5096546a52d72540ad0852cb975bf0e4e8876f1050e8240bc"} Oct 01 07:27:37 crc kubenswrapper[4837]: I1001 07:27:37.832642 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e16d452e-cf98-48c8-bde7-f003276bfa37" path="/var/lib/kubelet/pods/e16d452e-cf98-48c8-bde7-f003276bfa37/volumes" Oct 01 07:27:38 crc kubenswrapper[4837]: I1001 07:27:38.446256 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerStarted","Data":"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae"} Oct 01 07:27:38 crc kubenswrapper[4837]: I1001 07:27:38.446324 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerStarted","Data":"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57"} Oct 01 07:27:38 crc kubenswrapper[4837]: I1001 07:27:38.448458 4837 generic.go:334] "Generic (PLEG): container finished" podID="a1ca37a0-8611-42e6-b365-7ed2aa8dc811" containerID="8b0122a9f0b165396dc68834bcd670b9d796dd63c3cd41cd5cc6e187caf12ecf" exitCode=0 Oct 01 07:27:38 crc kubenswrapper[4837]: I1001 07:27:38.448523 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-l5h6t" event={"ID":"a1ca37a0-8611-42e6-b365-7ed2aa8dc811","Type":"ContainerDied","Data":"8b0122a9f0b165396dc68834bcd670b9d796dd63c3cd41cd5cc6e187caf12ecf"} Oct 01 07:27:38 crc kubenswrapper[4837]: I1001 07:27:38.490932 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.49067315 podStartE2EDuration="2.49067315s" podCreationTimestamp="2025-10-01 07:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:38.472646523 +0000 UTC m=+1315.314254008" watchObservedRunningTime="2025-10-01 07:27:38.49067315 +0000 UTC m=+1315.332280635" Oct 01 07:27:39 crc kubenswrapper[4837]: I1001 07:27:39.463550 4837 generic.go:334] "Generic (PLEG): container finished" podID="aef375f9-54f9-4c27-88c8-8a8345d4a5f8" containerID="8b0403847fa1e913e31eb2c874d46883eb7f9de350b911b8a7adf2ee314d096a" exitCode=0 Oct 01 07:27:39 crc kubenswrapper[4837]: I1001 07:27:39.463659 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" event={"ID":"aef375f9-54f9-4c27-88c8-8a8345d4a5f8","Type":"ContainerDied","Data":"8b0403847fa1e913e31eb2c874d46883eb7f9de350b911b8a7adf2ee314d096a"} Oct 01 07:27:39 crc kubenswrapper[4837]: I1001 07:27:39.969950 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.063403 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle\") pod \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.063538 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data\") pod \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.063566 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqvws\" (UniqueName: \"kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws\") pod \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.063723 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts\") pod \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\" (UID: \"a1ca37a0-8611-42e6-b365-7ed2aa8dc811\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.071927 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts" (OuterVolumeSpecName: "scripts") pod "a1ca37a0-8611-42e6-b365-7ed2aa8dc811" (UID: "a1ca37a0-8611-42e6-b365-7ed2aa8dc811"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.076054 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws" (OuterVolumeSpecName: "kube-api-access-sqvws") pod "a1ca37a0-8611-42e6-b365-7ed2aa8dc811" (UID: "a1ca37a0-8611-42e6-b365-7ed2aa8dc811"). InnerVolumeSpecName "kube-api-access-sqvws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.117776 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data" (OuterVolumeSpecName: "config-data") pod "a1ca37a0-8611-42e6-b365-7ed2aa8dc811" (UID: "a1ca37a0-8611-42e6-b365-7ed2aa8dc811"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.123073 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1ca37a0-8611-42e6-b365-7ed2aa8dc811" (UID: "a1ca37a0-8611-42e6-b365-7ed2aa8dc811"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.166932 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.166987 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.167003 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.167018 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqvws\" (UniqueName: \"kubernetes.io/projected/a1ca37a0-8611-42e6-b365-7ed2aa8dc811-kube-api-access-sqvws\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.494140 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-l5h6t" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.494661 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-l5h6t" event={"ID":"a1ca37a0-8611-42e6-b365-7ed2aa8dc811","Type":"ContainerDied","Data":"685b2ba643278077e23abd34c7ad6e6fff96f0cf27e068edbae666856d6e10f9"} Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.494735 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="685b2ba643278077e23abd34c7ad6e6fff96f0cf27e068edbae666856d6e10f9" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.728318 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.728581 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-log" containerID="cri-o://68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" gracePeriod=30 Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.728662 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-api" containerID="cri-o://33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" gracePeriod=30 Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.746744 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.746929 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="387b138c-ac22-411d-9692-41f7146df692" containerName="nova-scheduler-scheduler" containerID="cri-o://0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a" gracePeriod=30 Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.753996 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.754189 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-log" containerID="cri-o://d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" gracePeriod=30 Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.754319 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-metadata" containerID="cri-o://e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" gracePeriod=30 Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.823559 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.886276 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f76mg\" (UniqueName: \"kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg\") pod \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.886370 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle\") pod \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.886458 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data\") pod \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.886596 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts\") pod \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\" (UID: \"aef375f9-54f9-4c27-88c8-8a8345d4a5f8\") " Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.894247 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg" (OuterVolumeSpecName: "kube-api-access-f76mg") pod "aef375f9-54f9-4c27-88c8-8a8345d4a5f8" (UID: "aef375f9-54f9-4c27-88c8-8a8345d4a5f8"). InnerVolumeSpecName "kube-api-access-f76mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.894762 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts" (OuterVolumeSpecName: "scripts") pod "aef375f9-54f9-4c27-88c8-8a8345d4a5f8" (UID: "aef375f9-54f9-4c27-88c8-8a8345d4a5f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.916968 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aef375f9-54f9-4c27-88c8-8a8345d4a5f8" (UID: "aef375f9-54f9-4c27-88c8-8a8345d4a5f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.923908 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data" (OuterVolumeSpecName: "config-data") pod "aef375f9-54f9-4c27-88c8-8a8345d4a5f8" (UID: "aef375f9-54f9-4c27-88c8-8a8345d4a5f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.989426 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.989461 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.989471 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f76mg\" (UniqueName: \"kubernetes.io/projected/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-kube-api-access-f76mg\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:40 crc kubenswrapper[4837]: I1001 07:27:40.989481 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef375f9-54f9-4c27-88c8-8a8345d4a5f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.037985 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.114197 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.114461 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="dnsmasq-dns" containerID="cri-o://c5346b4bc62a9eb24a611ae5cf881437a77d4a6be9f5e881ba9a93a20c875511" gracePeriod=10 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.305373 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.307068 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395292 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs\") pod \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395403 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data\") pod \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395462 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs\") pod \"1258cc9c-10cb-4807-91de-e38be12f7533\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395520 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmfn4\" (UniqueName: \"kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4\") pod \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395543 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data\") pod \"1258cc9c-10cb-4807-91de-e38be12f7533\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395562 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle\") pod \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\" (UID: \"a5b6f5e1-7300-4744-b1e2-b6b7491297bd\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395579 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs\") pod \"1258cc9c-10cb-4807-91de-e38be12f7533\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395719 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjm89\" (UniqueName: \"kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89\") pod \"1258cc9c-10cb-4807-91de-e38be12f7533\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395742 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle\") pod \"1258cc9c-10cb-4807-91de-e38be12f7533\" (UID: \"1258cc9c-10cb-4807-91de-e38be12f7533\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.395811 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs" (OuterVolumeSpecName: "logs") pod "1258cc9c-10cb-4807-91de-e38be12f7533" (UID: "1258cc9c-10cb-4807-91de-e38be12f7533"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.396087 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs" (OuterVolumeSpecName: "logs") pod "a5b6f5e1-7300-4744-b1e2-b6b7491297bd" (UID: "a5b6f5e1-7300-4744-b1e2-b6b7491297bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.396142 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1258cc9c-10cb-4807-91de-e38be12f7533-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.400833 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89" (OuterVolumeSpecName: "kube-api-access-zjm89") pod "1258cc9c-10cb-4807-91de-e38be12f7533" (UID: "1258cc9c-10cb-4807-91de-e38be12f7533"). InnerVolumeSpecName "kube-api-access-zjm89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.401745 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4" (OuterVolumeSpecName: "kube-api-access-vmfn4") pod "a5b6f5e1-7300-4744-b1e2-b6b7491297bd" (UID: "a5b6f5e1-7300-4744-b1e2-b6b7491297bd"). InnerVolumeSpecName "kube-api-access-vmfn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.440640 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1258cc9c-10cb-4807-91de-e38be12f7533" (UID: "1258cc9c-10cb-4807-91de-e38be12f7533"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.448704 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5b6f5e1-7300-4744-b1e2-b6b7491297bd" (UID: "a5b6f5e1-7300-4744-b1e2-b6b7491297bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.451241 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data" (OuterVolumeSpecName: "config-data") pod "a5b6f5e1-7300-4744-b1e2-b6b7491297bd" (UID: "a5b6f5e1-7300-4744-b1e2-b6b7491297bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.458923 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data" (OuterVolumeSpecName: "config-data") pod "1258cc9c-10cb-4807-91de-e38be12f7533" (UID: "1258cc9c-10cb-4807-91de-e38be12f7533"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.465008 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1258cc9c-10cb-4807-91de-e38be12f7533" (UID: "1258cc9c-10cb-4807-91de-e38be12f7533"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497572 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497605 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497618 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmfn4\" (UniqueName: \"kubernetes.io/projected/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-kube-api-access-vmfn4\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497628 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497637 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5b6f5e1-7300-4744-b1e2-b6b7491297bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497647 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497655 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjm89\" (UniqueName: \"kubernetes.io/projected/1258cc9c-10cb-4807-91de-e38be12f7533-kube-api-access-zjm89\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.497663 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1258cc9c-10cb-4807-91de-e38be12f7533-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.512982 4837 generic.go:334] "Generic (PLEG): container finished" podID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerID="c5346b4bc62a9eb24a611ae5cf881437a77d4a6be9f5e881ba9a93a20c875511" exitCode=0 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.513046 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" event={"ID":"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2","Type":"ContainerDied","Data":"c5346b4bc62a9eb24a611ae5cf881437a77d4a6be9f5e881ba9a93a20c875511"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.515884 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.515885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerDied","Data":"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.515949 4837 scope.go:117] "RemoveContainer" containerID="e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.515651 4837 generic.go:334] "Generic (PLEG): container finished" podID="1258cc9c-10cb-4807-91de-e38be12f7533" containerID="e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" exitCode=0 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.516026 4837 generic.go:334] "Generic (PLEG): container finished" podID="1258cc9c-10cb-4807-91de-e38be12f7533" containerID="d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" exitCode=143 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.520825 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerDied","Data":"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.520920 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1258cc9c-10cb-4807-91de-e38be12f7533","Type":"ContainerDied","Data":"00b642ab103d15b5096546a52d72540ad0852cb975bf0e4e8876f1050e8240bc"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529543 4837 generic.go:334] "Generic (PLEG): container finished" podID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerID="33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" exitCode=0 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529569 4837 generic.go:334] "Generic (PLEG): container finished" podID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerID="68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" exitCode=143 Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529605 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerDied","Data":"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerDied","Data":"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529637 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a5b6f5e1-7300-4744-b1e2-b6b7491297bd","Type":"ContainerDied","Data":"d72640bab6def49bbe8c0acb5688295c9f2fd5b4f6eb4ab3aaa91e6762572ed5"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.529712 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.551253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" event={"ID":"aef375f9-54f9-4c27-88c8-8a8345d4a5f8","Type":"ContainerDied","Data":"e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17"} Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.551292 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e34c78138a24fa6327ebb7d674a5a9facb0d43e880ee01dadd73de40294adb17" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.551359 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lrdgf" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.581525 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.591456 4837 scope.go:117] "RemoveContainer" containerID="d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.592869 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596561 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-log" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596588 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-log" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596599 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef375f9-54f9-4c27-88c8-8a8345d4a5f8" containerName="nova-cell1-conductor-db-sync" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596606 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef375f9-54f9-4c27-88c8-8a8345d4a5f8" containerName="nova-cell1-conductor-db-sync" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596644 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ca37a0-8611-42e6-b365-7ed2aa8dc811" containerName="nova-manage" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596653 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ca37a0-8611-42e6-b365-7ed2aa8dc811" containerName="nova-manage" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596678 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-log" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596814 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-log" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596831 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-api" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596838 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-api" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.596846 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-metadata" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.596852 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-metadata" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597077 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ca37a0-8611-42e6-b365-7ed2aa8dc811" containerName="nova-manage" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597089 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-log" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597122 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-log" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597135 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" containerName="nova-metadata-metadata" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597151 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef375f9-54f9-4c27-88c8-8a8345d4a5f8" containerName="nova-cell1-conductor-db-sync" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.597158 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" containerName="nova-api-api" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.606091 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.618064 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.626950 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.633953 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.646029 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.646843 4837 scope.go:117] "RemoveContainer" containerID="e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.647632 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.654490 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae\": container with ID starting with e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae not found: ID does not exist" containerID="e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.654550 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae"} err="failed to get container status \"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae\": rpc error: code = NotFound desc = could not find container \"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae\": container with ID starting with e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.654578 4837 scope.go:117] "RemoveContainer" containerID="d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.654605 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.657329 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57\": container with ID starting with d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57 not found: ID does not exist" containerID="d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.657388 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57"} err="failed to get container status \"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57\": rpc error: code = NotFound desc = could not find container \"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57\": container with ID starting with d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57 not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.657414 4837 scope.go:117] "RemoveContainer" containerID="e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.658630 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae"} err="failed to get container status \"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae\": rpc error: code = NotFound desc = could not find container \"e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae\": container with ID starting with e591c69938f9ffff3b13ba2144fc6f8e9f5e76df1ecc0770f4e280f8e8599eae not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.658651 4837 scope.go:117] "RemoveContainer" containerID="d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.658731 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.659043 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57"} err="failed to get container status \"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57\": rpc error: code = NotFound desc = could not find container \"d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57\": container with ID starting with d5ce4253de53afce8c2e6b1ce696ef7a3d1eaa69e81938fd10665e4f56e87e57 not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.659062 4837 scope.go:117] "RemoveContainer" containerID="33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.666864 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.681425 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.684973 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.685796 4837 scope.go:117] "RemoveContainer" containerID="68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.689017 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.689313 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="dnsmasq-dns" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.689329 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="dnsmasq-dns" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.689359 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="init" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.689366 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="init" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.689527 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" containerName="dnsmasq-dns" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.690452 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.694727 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.694929 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.703354 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.703914 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704006 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704038 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704086 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704260 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z8n7\" (UniqueName: \"kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704283 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config\") pod \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\" (UID: \"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2\") " Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704507 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t962t\" (UniqueName: \"kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.704602 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vtn5\" (UniqueName: \"kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.705351 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.705381 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.705423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.705443 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.710302 4837 scope.go:117] "RemoveContainer" containerID="33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.710661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7" (OuterVolumeSpecName: "kube-api-access-9z8n7") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "kube-api-access-9z8n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.711186 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643\": container with ID starting with 33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643 not found: ID does not exist" containerID="33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.711220 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643"} err="failed to get container status \"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643\": rpc error: code = NotFound desc = could not find container \"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643\": container with ID starting with 33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643 not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.711243 4837 scope.go:117] "RemoveContainer" containerID="68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" Oct 01 07:27:41 crc kubenswrapper[4837]: E1001 07:27:41.719258 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad\": container with ID starting with 68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad not found: ID does not exist" containerID="68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.719416 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad"} err="failed to get container status \"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad\": rpc error: code = NotFound desc = could not find container \"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad\": container with ID starting with 68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.719509 4837 scope.go:117] "RemoveContainer" containerID="33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.720516 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643"} err="failed to get container status \"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643\": rpc error: code = NotFound desc = could not find container \"33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643\": container with ID starting with 33ffb864ea27805447faceff954ab04b4aaf8fb176fcd49e5e1a17a3e05cb643 not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.720559 4837 scope.go:117] "RemoveContainer" containerID="68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.720814 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad"} err="failed to get container status \"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad\": rpc error: code = NotFound desc = could not find container \"68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad\": container with ID starting with 68b5a85848451cc24009feffcfd010af47c5a74f3bc954f8e105c360e363bdad not found: ID does not exist" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.771641 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.773474 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.778236 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config" (OuterVolumeSpecName: "config") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.787508 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.796293 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" (UID: "4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810258 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vtn5\" (UniqueName: \"kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810330 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8p69\" (UniqueName: \"kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810362 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810401 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810429 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810460 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810479 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810495 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810524 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t962t\" (UniqueName: \"kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810551 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810571 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810631 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z8n7\" (UniqueName: \"kubernetes.io/projected/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-kube-api-access-9z8n7\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810642 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810652 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810662 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810670 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.810679 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.811276 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.814557 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.819089 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.820061 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.821362 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.826186 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1258cc9c-10cb-4807-91de-e38be12f7533" path="/var/lib/kubelet/pods/1258cc9c-10cb-4807-91de-e38be12f7533/volumes" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.826793 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5b6f5e1-7300-4744-b1e2-b6b7491297bd" path="/var/lib/kubelet/pods/a5b6f5e1-7300-4744-b1e2-b6b7491297bd/volumes" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.829357 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vtn5\" (UniqueName: \"kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5\") pod \"nova-api-0\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " pod="openstack/nova-api-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.835516 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t962t\" (UniqueName: \"kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t\") pod \"nova-cell1-conductor-0\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.911325 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.911501 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8p69\" (UniqueName: \"kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.911563 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.911608 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.911654 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.920647 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.920923 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.922444 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.923347 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.959321 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8p69\" (UniqueName: \"kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69\") pod \"nova-metadata-0\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " pod="openstack/nova-metadata-0" Oct 01 07:27:41 crc kubenswrapper[4837]: I1001 07:27:41.983441 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.009577 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.028113 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.243006 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.360433 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle\") pod \"387b138c-ac22-411d-9692-41f7146df692\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.360543 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data\") pod \"387b138c-ac22-411d-9692-41f7146df692\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.360642 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqv55\" (UniqueName: \"kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55\") pod \"387b138c-ac22-411d-9692-41f7146df692\" (UID: \"387b138c-ac22-411d-9692-41f7146df692\") " Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.368569 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55" (OuterVolumeSpecName: "kube-api-access-lqv55") pod "387b138c-ac22-411d-9692-41f7146df692" (UID: "387b138c-ac22-411d-9692-41f7146df692"). InnerVolumeSpecName "kube-api-access-lqv55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.396636 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data" (OuterVolumeSpecName: "config-data") pod "387b138c-ac22-411d-9692-41f7146df692" (UID: "387b138c-ac22-411d-9692-41f7146df692"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.401293 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "387b138c-ac22-411d-9692-41f7146df692" (UID: "387b138c-ac22-411d-9692-41f7146df692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.462152 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqv55\" (UniqueName: \"kubernetes.io/projected/387b138c-ac22-411d-9692-41f7146df692-kube-api-access-lqv55\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.462214 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.462226 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387b138c-ac22-411d-9692-41f7146df692-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.542554 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: W1001 07:27:42.544349 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc46a0ffe_6c41_4d34_bbe9_8db505bbe46b.slice/crio-276a680b19af91fda764eff2425453cc9b6c441eae10dd1481ff673a9cb923ad WatchSource:0}: Error finding container 276a680b19af91fda764eff2425453cc9b6c441eae10dd1481ff673a9cb923ad: Status 404 returned error can't find the container with id 276a680b19af91fda764eff2425453cc9b6c441eae10dd1481ff673a9cb923ad Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.561119 4837 generic.go:334] "Generic (PLEG): container finished" podID="387b138c-ac22-411d-9692-41f7146df692" containerID="0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a" exitCode=0 Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.561215 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"387b138c-ac22-411d-9692-41f7146df692","Type":"ContainerDied","Data":"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a"} Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.561249 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"387b138c-ac22-411d-9692-41f7146df692","Type":"ContainerDied","Data":"d58bc394415238a254ccf79511faf5c6019ad590722b2273c23e8ee6ab28e817"} Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.561277 4837 scope.go:117] "RemoveContainer" containerID="0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.561264 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.566232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" event={"ID":"4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2","Type":"ContainerDied","Data":"8bf08a7d2853d77f5f609a6b3474f33a557bc5e3ed700d5f9587c332e4276bc7"} Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.566302 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6797bbc6b9-2p4ll" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.567530 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b","Type":"ContainerStarted","Data":"276a680b19af91fda764eff2425453cc9b6c441eae10dd1481ff673a9cb923ad"} Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.589765 4837 scope.go:117] "RemoveContainer" containerID="0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a" Oct 01 07:27:42 crc kubenswrapper[4837]: E1001 07:27:42.590459 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a\": container with ID starting with 0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a not found: ID does not exist" containerID="0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.590503 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a"} err="failed to get container status \"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a\": rpc error: code = NotFound desc = could not find container \"0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a\": container with ID starting with 0275a15d5332c7aca84ef997505597c9b9caf2b3878a2dbc30bd41c0eb8ff67a not found: ID does not exist" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.590533 4837 scope.go:117] "RemoveContainer" containerID="c5346b4bc62a9eb24a611ae5cf881437a77d4a6be9f5e881ba9a93a20c875511" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.595921 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.616661 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6797bbc6b9-2p4ll"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.636427 4837 scope.go:117] "RemoveContainer" containerID="56bc8d2d9f6ccd18df163e5e792c85f317dd163e845422f27facf9060502b2b5" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.638188 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.651664 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.663514 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: E1001 07:27:42.664079 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387b138c-ac22-411d-9692-41f7146df692" containerName="nova-scheduler-scheduler" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.664099 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="387b138c-ac22-411d-9692-41f7146df692" containerName="nova-scheduler-scheduler" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.664347 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="387b138c-ac22-411d-9692-41f7146df692" containerName="nova-scheduler-scheduler" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.665213 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.668161 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 07:27:42 crc kubenswrapper[4837]: W1001 07:27:42.669767 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod981c4ae4_7e11_4a52_a559_2ebb55c622d5.slice/crio-37d31dc9a7e461c65693d223e45838b76ebe19159df9b21662019a09caa515f6 WatchSource:0}: Error finding container 37d31dc9a7e461c65693d223e45838b76ebe19159df9b21662019a09caa515f6: Status 404 returned error can't find the container with id 37d31dc9a7e461c65693d223e45838b76ebe19159df9b21662019a09caa515f6 Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.673451 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.687490 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.729506 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:27:42 crc kubenswrapper[4837]: W1001 07:27:42.729592 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3e952ac_c2ad_4b22_9eab_19c5f36f703c.slice/crio-a697c9abc111e268e4beba0c4f657072bde1cba810dff254aac8c36f92e79835 WatchSource:0}: Error finding container a697c9abc111e268e4beba0c4f657072bde1cba810dff254aac8c36f92e79835: Status 404 returned error can't find the container with id a697c9abc111e268e4beba0c4f657072bde1cba810dff254aac8c36f92e79835 Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.769412 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.770209 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68758\" (UniqueName: \"kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.770261 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.872207 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68758\" (UniqueName: \"kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.872325 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.872831 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.878163 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.878999 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:42 crc kubenswrapper[4837]: I1001 07:27:42.892833 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68758\" (UniqueName: \"kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758\") pod \"nova-scheduler-0\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " pod="openstack/nova-scheduler-0" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.014316 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:27:43 crc kubenswrapper[4837]: W1001 07:27:43.304914 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83caf5e4_336c_4def_9eb6_ba7ad607cfb2.slice/crio-27aad9ded82ee9c5febd8f30d69e6025ee77e885614a14d7fc6112981b263a70 WatchSource:0}: Error finding container 27aad9ded82ee9c5febd8f30d69e6025ee77e885614a14d7fc6112981b263a70: Status 404 returned error can't find the container with id 27aad9ded82ee9c5febd8f30d69e6025ee77e885614a14d7fc6112981b263a70 Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.308594 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.580360 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83caf5e4-336c-4def-9eb6-ba7ad607cfb2","Type":"ContainerStarted","Data":"589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.580715 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83caf5e4-336c-4def-9eb6-ba7ad607cfb2","Type":"ContainerStarted","Data":"27aad9ded82ee9c5febd8f30d69e6025ee77e885614a14d7fc6112981b263a70"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.585639 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerStarted","Data":"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.585724 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerStarted","Data":"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.585738 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerStarted","Data":"37d31dc9a7e461c65693d223e45838b76ebe19159df9b21662019a09caa515f6"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.591750 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerStarted","Data":"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.591783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerStarted","Data":"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.591792 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerStarted","Data":"a697c9abc111e268e4beba0c4f657072bde1cba810dff254aac8c36f92e79835"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.596538 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b","Type":"ContainerStarted","Data":"ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901"} Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.596704 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.608524 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.6085066019999998 podStartE2EDuration="1.608506602s" podCreationTimestamp="2025-10-01 07:27:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:43.59551591 +0000 UTC m=+1320.437123375" watchObservedRunningTime="2025-10-01 07:27:43.608506602 +0000 UTC m=+1320.450114057" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.626810 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.626788486 podStartE2EDuration="2.626788486s" podCreationTimestamp="2025-10-01 07:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:43.619040023 +0000 UTC m=+1320.460647488" watchObservedRunningTime="2025-10-01 07:27:43.626788486 +0000 UTC m=+1320.468395931" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.643226 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.643209894 podStartE2EDuration="2.643209894s" podCreationTimestamp="2025-10-01 07:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:43.63944349 +0000 UTC m=+1320.481050975" watchObservedRunningTime="2025-10-01 07:27:43.643209894 +0000 UTC m=+1320.484817349" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.671887 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.671857215 podStartE2EDuration="2.671857215s" podCreationTimestamp="2025-10-01 07:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:27:43.660308159 +0000 UTC m=+1320.501915614" watchObservedRunningTime="2025-10-01 07:27:43.671857215 +0000 UTC m=+1320.513464700" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.831391 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387b138c-ac22-411d-9692-41f7146df692" path="/var/lib/kubelet/pods/387b138c-ac22-411d-9692-41f7146df692/volumes" Oct 01 07:27:43 crc kubenswrapper[4837]: I1001 07:27:43.832468 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2" path="/var/lib/kubelet/pods/4b1a5fba-2c5e-4f3e-9b96-0ad4629acdb2/volumes" Oct 01 07:27:45 crc kubenswrapper[4837]: I1001 07:27:45.754435 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 07:27:47 crc kubenswrapper[4837]: I1001 07:27:47.028767 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 07:27:47 crc kubenswrapper[4837]: I1001 07:27:47.029096 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 07:27:48 crc kubenswrapper[4837]: I1001 07:27:48.014789 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.334832 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.335021 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="14422052-034a-4e59-a29b-6be7aa9d077b" containerName="kube-state-metrics" containerID="cri-o://6a5050452166b7757eb54854930552e444559ebd525aa9ed438a4a78f9e85d22" gracePeriod=30 Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.668949 4837 generic.go:334] "Generic (PLEG): container finished" podID="14422052-034a-4e59-a29b-6be7aa9d077b" containerID="6a5050452166b7757eb54854930552e444559ebd525aa9ed438a4a78f9e85d22" exitCode=2 Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.669058 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"14422052-034a-4e59-a29b-6be7aa9d077b","Type":"ContainerDied","Data":"6a5050452166b7757eb54854930552e444559ebd525aa9ed438a4a78f9e85d22"} Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.812045 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.912271 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q76hj\" (UniqueName: \"kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj\") pod \"14422052-034a-4e59-a29b-6be7aa9d077b\" (UID: \"14422052-034a-4e59-a29b-6be7aa9d077b\") " Oct 01 07:27:49 crc kubenswrapper[4837]: I1001 07:27:49.918093 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj" (OuterVolumeSpecName: "kube-api-access-q76hj") pod "14422052-034a-4e59-a29b-6be7aa9d077b" (UID: "14422052-034a-4e59-a29b-6be7aa9d077b"). InnerVolumeSpecName "kube-api-access-q76hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.015176 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q76hj\" (UniqueName: \"kubernetes.io/projected/14422052-034a-4e59-a29b-6be7aa9d077b-kube-api-access-q76hj\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.685043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"14422052-034a-4e59-a29b-6be7aa9d077b","Type":"ContainerDied","Data":"fd4306543ebd110b5209a1acde5f8b9d81a4e18a43c3efb841eb4f85d0ac3f70"} Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.685097 4837 scope.go:117] "RemoveContainer" containerID="6a5050452166b7757eb54854930552e444559ebd525aa9ed438a4a78f9e85d22" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.685228 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.726427 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.737855 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.747860 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:50 crc kubenswrapper[4837]: E1001 07:27:50.748369 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14422052-034a-4e59-a29b-6be7aa9d077b" containerName="kube-state-metrics" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.748393 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="14422052-034a-4e59-a29b-6be7aa9d077b" containerName="kube-state-metrics" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.748633 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="14422052-034a-4e59-a29b-6be7aa9d077b" containerName="kube-state-metrics" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.749388 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.755130 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.755212 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.765276 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.829438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.829577 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.829670 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.829757 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6flkz\" (UniqueName: \"kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.930748 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.930854 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.930918 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6flkz\" (UniqueName: \"kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.930987 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.935674 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.936453 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.938371 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:50 crc kubenswrapper[4837]: I1001 07:27:50.956802 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6flkz\" (UniqueName: \"kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz\") pod \"kube-state-metrics-0\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " pod="openstack/kube-state-metrics-0" Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.031234 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.031808 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-central-agent" containerID="cri-o://8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e" gracePeriod=30 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.032229 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="proxy-httpd" containerID="cri-o://5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33" gracePeriod=30 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.032285 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="sg-core" containerID="cri-o://41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705" gracePeriod=30 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.032336 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-notification-agent" containerID="cri-o://37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6" gracePeriod=30 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.071329 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.554747 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:27:51 crc kubenswrapper[4837]: W1001 07:27:51.557192 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2851153_f304_4041_9de2_045c0d92526d.slice/crio-47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f WatchSource:0}: Error finding container 47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f: Status 404 returned error can't find the container with id 47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708658 4837 generic.go:334] "Generic (PLEG): container finished" podID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerID="5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33" exitCode=0 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708705 4837 generic.go:334] "Generic (PLEG): container finished" podID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerID="41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705" exitCode=2 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708714 4837 generic.go:334] "Generic (PLEG): container finished" podID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerID="8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e" exitCode=0 Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708761 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerDied","Data":"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33"} Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708858 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerDied","Data":"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705"} Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.708896 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerDied","Data":"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e"} Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.713061 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2851153-f304-4041-9de2-045c0d92526d","Type":"ContainerStarted","Data":"47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f"} Oct 01 07:27:51 crc kubenswrapper[4837]: I1001 07:27:51.837551 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14422052-034a-4e59-a29b-6be7aa9d077b" path="/var/lib/kubelet/pods/14422052-034a-4e59-a29b-6be7aa9d077b/volumes" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.010563 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.010640 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.028804 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.028872 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.049496 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.727521 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2851153-f304-4041-9de2-045c0d92526d","Type":"ContainerStarted","Data":"034af4cc06f98a5a1c1bf69b7c633186fde48f42cccb9b972d748a605cd9eec8"} Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.728490 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 07:27:52 crc kubenswrapper[4837]: I1001 07:27:52.749318 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.392762354 podStartE2EDuration="2.749301888s" podCreationTimestamp="2025-10-01 07:27:50 +0000 UTC" firstStartedPulling="2025-10-01 07:27:51.560097047 +0000 UTC m=+1328.401704502" lastFinishedPulling="2025-10-01 07:27:51.916636571 +0000 UTC m=+1328.758244036" observedRunningTime="2025-10-01 07:27:52.74656215 +0000 UTC m=+1329.588169605" watchObservedRunningTime="2025-10-01 07:27:52.749301888 +0000 UTC m=+1329.590909343" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.015610 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.052997 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.105825 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.105839 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.105885 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.105920 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.106112 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.106592 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.106625 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.107415 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.107499 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92" gracePeriod=600 Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.385794 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.573816 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574142 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp962\" (UniqueName: \"kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574256 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574366 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574411 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574451 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd\") pod \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\" (UID: \"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7\") " Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.574840 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.581473 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.584400 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts" (OuterVolumeSpecName: "scripts") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.592280 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962" (OuterVolumeSpecName: "kube-api-access-zp962") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "kube-api-access-zp962". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.621121 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.669316 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677425 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677464 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677477 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677489 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677501 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp962\" (UniqueName: \"kubernetes.io/projected/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-kube-api-access-zp962\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.677517 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.694744 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data" (OuterVolumeSpecName: "config-data") pod "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" (UID: "a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.737746 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92" exitCode=0 Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.737812 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92"} Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.737870 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d"} Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.738228 4837 scope.go:117] "RemoveContainer" containerID="992e57ff761899956e03a4f303f8ef8b0498eb9d9618228358bba77b87630e86" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.740966 4837 generic.go:334] "Generic (PLEG): container finished" podID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerID="37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6" exitCode=0 Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.741927 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.742124 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerDied","Data":"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6"} Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.742183 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7","Type":"ContainerDied","Data":"3acd5b7fa9c5da9e99678f4a8ec86451cdd10b893899312f908c083b4c216abf"} Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.782365 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.787891 4837 scope.go:117] "RemoveContainer" containerID="5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.790362 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.801479 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.812486 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.812839 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-notification-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.812855 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-notification-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.812870 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="proxy-httpd" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.812878 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="proxy-httpd" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.812895 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="sg-core" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.812901 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="sg-core" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.812914 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-central-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.812920 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-central-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.813094 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="sg-core" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.813105 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="proxy-httpd" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.813117 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-central-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.813330 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" containerName="ceilometer-notification-agent" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.817632 4837 scope.go:117] "RemoveContainer" containerID="41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.823172 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.823498 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.826034 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.826258 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.826382 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.846892 4837 scope.go:117] "RemoveContainer" containerID="37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.849467 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7" path="/var/lib/kubelet/pods/a11fd276-c6d2-4073-a9f5-a91e7eb1cfb7/volumes" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.852268 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.878208 4837 scope.go:117] "RemoveContainer" containerID="8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.904788 4837 scope.go:117] "RemoveContainer" containerID="5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.905212 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33\": container with ID starting with 5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33 not found: ID does not exist" containerID="5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905243 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33"} err="failed to get container status \"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33\": rpc error: code = NotFound desc = could not find container \"5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33\": container with ID starting with 5c280a9600bf94db5eba754e042ffe1060b0876c7460dd4802e709bc079a8a33 not found: ID does not exist" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905265 4837 scope.go:117] "RemoveContainer" containerID="41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.905577 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705\": container with ID starting with 41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705 not found: ID does not exist" containerID="41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905612 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705"} err="failed to get container status \"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705\": rpc error: code = NotFound desc = could not find container \"41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705\": container with ID starting with 41a9f1d7b0035760f780324e2e856f6b9c2c5d62bb7a513fe3f647e82a6fc705 not found: ID does not exist" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905624 4837 scope.go:117] "RemoveContainer" containerID="37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.905868 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6\": container with ID starting with 37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6 not found: ID does not exist" containerID="37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905889 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6"} err="failed to get container status \"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6\": rpc error: code = NotFound desc = could not find container \"37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6\": container with ID starting with 37cf337ab0580452afd408f7258eac3ed1c76abdaec2596688208875607053d6 not found: ID does not exist" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.905902 4837 scope.go:117] "RemoveContainer" containerID="8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e" Oct 01 07:27:53 crc kubenswrapper[4837]: E1001 07:27:53.906119 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e\": container with ID starting with 8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e not found: ID does not exist" containerID="8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e" Oct 01 07:27:53 crc kubenswrapper[4837]: I1001 07:27:53.906136 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e"} err="failed to get container status \"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e\": rpc error: code = NotFound desc = could not find container \"8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e\": container with ID starting with 8c9bf083200568bdf22e0ab6e4571ef89b7e4466b5ea154066855e74c5a4470e not found: ID does not exist" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.007459 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.007780 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.008559 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.008606 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.008649 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.008837 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmm5f\" (UniqueName: \"kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.008958 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.009006 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.110478 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmm5f\" (UniqueName: \"kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.110683 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.110822 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.110990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111116 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111241 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111380 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111472 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111611 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.111643 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.115897 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.115957 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.118092 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.118169 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.125977 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmm5f\" (UniqueName: \"kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.128873 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data\") pod \"ceilometer-0\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.158530 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.636260 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:27:54 crc kubenswrapper[4837]: I1001 07:27:54.753419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerStarted","Data":"aa16baa5fb2078db2c6f845f255f05b3a19112a84d14e15a2ff502aec9a8ce54"} Oct 01 07:27:56 crc kubenswrapper[4837]: I1001 07:27:56.788185 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerStarted","Data":"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b"} Oct 01 07:27:56 crc kubenswrapper[4837]: I1001 07:27:56.788894 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerStarted","Data":"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19"} Oct 01 07:27:57 crc kubenswrapper[4837]: I1001 07:27:57.806042 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerStarted","Data":"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333"} Oct 01 07:27:59 crc kubenswrapper[4837]: I1001 07:27:59.830905 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerStarted","Data":"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f"} Oct 01 07:27:59 crc kubenswrapper[4837]: I1001 07:27:59.832118 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:27:59 crc kubenswrapper[4837]: I1001 07:27:59.860180 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.451080013 podStartE2EDuration="6.860159234s" podCreationTimestamp="2025-10-01 07:27:53 +0000 UTC" firstStartedPulling="2025-10-01 07:27:54.640450402 +0000 UTC m=+1331.482057857" lastFinishedPulling="2025-10-01 07:27:59.049529583 +0000 UTC m=+1335.891137078" observedRunningTime="2025-10-01 07:27:59.855477208 +0000 UTC m=+1336.697084673" watchObservedRunningTime="2025-10-01 07:27:59.860159234 +0000 UTC m=+1336.701766689" Oct 01 07:28:01 crc kubenswrapper[4837]: I1001 07:28:01.089131 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.017498 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.018785 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.018935 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.027338 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.046142 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.048300 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.052274 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.866853 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.874649 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 07:28:02 crc kubenswrapper[4837]: I1001 07:28:02.880157 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.097402 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.101638 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.129409 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.212663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.212749 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.212882 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwwxj\" (UniqueName: \"kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.212940 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.212980 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.213226 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.314576 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.314808 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.314918 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwwxj\" (UniqueName: \"kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.314995 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.315071 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.315189 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.316370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.317055 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.317600 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.318195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.318769 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.345179 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwwxj\" (UniqueName: \"kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj\") pod \"dnsmasq-dns-df6b7c479-f24qr\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.430626 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:03 crc kubenswrapper[4837]: I1001 07:28:03.910286 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:28:04 crc kubenswrapper[4837]: I1001 07:28:04.883680 4837 generic.go:334] "Generic (PLEG): container finished" podID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerID="9d47745a8b132f2982a19fd66ee8dfb226b79b906c64bbaffadbf5953e32dcb4" exitCode=0 Oct 01 07:28:04 crc kubenswrapper[4837]: I1001 07:28:04.883905 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" event={"ID":"34a848f4-7b85-4463-bdae-f763e6f9e7e5","Type":"ContainerDied","Data":"9d47745a8b132f2982a19fd66ee8dfb226b79b906c64bbaffadbf5953e32dcb4"} Oct 01 07:28:04 crc kubenswrapper[4837]: I1001 07:28:04.884035 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" event={"ID":"34a848f4-7b85-4463-bdae-f763e6f9e7e5","Type":"ContainerStarted","Data":"37138bfce86319542c491e929d2e04e41ee31e25ba2a9d815395bd4458116854"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.530138 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.530624 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-central-agent" containerID="cri-o://a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.530880 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="proxy-httpd" containerID="cri-o://1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.531061 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="sg-core" containerID="cri-o://03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.531073 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-notification-agent" containerID="cri-o://43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.831152 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.893618 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" event={"ID":"34a848f4-7b85-4463-bdae-f763e6f9e7e5","Type":"ContainerStarted","Data":"b5ac45277657c7bc51337c63b0014ecaf12e5118db0d75abef1c824653bddcaa"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.893744 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.896908 4837 generic.go:334] "Generic (PLEG): container finished" podID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerID="1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" exitCode=0 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.896938 4837 generic.go:334] "Generic (PLEG): container finished" podID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerID="03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" exitCode=2 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.896990 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerDied","Data":"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.897028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerDied","Data":"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898715 4837 generic.go:334] "Generic (PLEG): container finished" podID="b82e1abf-77fe-498c-8a09-2e509a551607" containerID="7e1bae08709298165b16e91922f8d331270e19bce0e857867200a431bc764409" exitCode=137 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898748 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b82e1abf-77fe-498c-8a09-2e509a551607","Type":"ContainerDied","Data":"7e1bae08709298165b16e91922f8d331270e19bce0e857867200a431bc764409"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898800 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b82e1abf-77fe-498c-8a09-2e509a551607","Type":"ContainerDied","Data":"11dd6ec3ed9f649bf0d5836beffb32c9ceec3b315ccc784b55881dc75e8bac14"} Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898816 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11dd6ec3ed9f649bf0d5836beffb32c9ceec3b315ccc784b55881dc75e8bac14" Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898902 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-log" containerID="cri-o://deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.898999 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-api" containerID="cri-o://7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a" gracePeriod=30 Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.920477 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" podStartSLOduration=2.92046003 podStartE2EDuration="2.92046003s" podCreationTimestamp="2025-10-01 07:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:05.913684422 +0000 UTC m=+1342.755291877" watchObservedRunningTime="2025-10-01 07:28:05.92046003 +0000 UTC m=+1342.762067485" Oct 01 07:28:05 crc kubenswrapper[4837]: I1001 07:28:05.929372 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.080866 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle\") pod \"b82e1abf-77fe-498c-8a09-2e509a551607\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.080905 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8655\" (UniqueName: \"kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655\") pod \"b82e1abf-77fe-498c-8a09-2e509a551607\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.081071 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data\") pod \"b82e1abf-77fe-498c-8a09-2e509a551607\" (UID: \"b82e1abf-77fe-498c-8a09-2e509a551607\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.087467 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655" (OuterVolumeSpecName: "kube-api-access-l8655") pod "b82e1abf-77fe-498c-8a09-2e509a551607" (UID: "b82e1abf-77fe-498c-8a09-2e509a551607"). InnerVolumeSpecName "kube-api-access-l8655". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.105006 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data" (OuterVolumeSpecName: "config-data") pod "b82e1abf-77fe-498c-8a09-2e509a551607" (UID: "b82e1abf-77fe-498c-8a09-2e509a551607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.106142 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b82e1abf-77fe-498c-8a09-2e509a551607" (UID: "b82e1abf-77fe-498c-8a09-2e509a551607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.186942 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.186974 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8655\" (UniqueName: \"kubernetes.io/projected/b82e1abf-77fe-498c-8a09-2e509a551607-kube-api-access-l8655\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.186987 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b82e1abf-77fe-498c-8a09-2e509a551607-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.420242 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.509104 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.509546 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.509674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.509814 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.509925 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.510040 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.510152 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmm5f\" (UniqueName: \"kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.510309 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts\") pod \"46d312f6-5f1b-4878-ad05-7146f85b77b4\" (UID: \"46d312f6-5f1b-4878-ad05-7146f85b77b4\") " Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.510284 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.510468 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.514841 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f" (OuterVolumeSpecName: "kube-api-access-nmm5f") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "kube-api-access-nmm5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.516229 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts" (OuterVolumeSpecName: "scripts") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.547041 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.597939 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612640 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612672 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612681 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46d312f6-5f1b-4878-ad05-7146f85b77b4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612703 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612714 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmm5f\" (UniqueName: \"kubernetes.io/projected/46d312f6-5f1b-4878-ad05-7146f85b77b4-kube-api-access-nmm5f\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.612721 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.613873 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.637754 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data" (OuterVolumeSpecName: "config-data") pod "46d312f6-5f1b-4878-ad05-7146f85b77b4" (UID: "46d312f6-5f1b-4878-ad05-7146f85b77b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.714729 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.714971 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46d312f6-5f1b-4878-ad05-7146f85b77b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912424 4837 generic.go:334] "Generic (PLEG): container finished" podID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerID="43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" exitCode=0 Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912468 4837 generic.go:334] "Generic (PLEG): container finished" podID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerID="a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" exitCode=0 Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912525 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerDied","Data":"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b"} Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912562 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerDied","Data":"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19"} Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912581 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46d312f6-5f1b-4878-ad05-7146f85b77b4","Type":"ContainerDied","Data":"aa16baa5fb2078db2c6f845f255f05b3a19112a84d14e15a2ff502aec9a8ce54"} Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912603 4837 scope.go:117] "RemoveContainer" containerID="1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.912797 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.922099 4837 generic.go:334] "Generic (PLEG): container finished" podID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerID="deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c" exitCode=143 Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.922147 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerDied","Data":"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c"} Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.922645 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.946985 4837 scope.go:117] "RemoveContainer" containerID="03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.967958 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.978965 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:06 crc kubenswrapper[4837]: I1001 07:28:06.982390 4837 scope.go:117] "RemoveContainer" containerID="43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.004479 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.018725 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.026540 4837 scope.go:117] "RemoveContainer" containerID="a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034183 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.034608 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b82e1abf-77fe-498c-8a09-2e509a551607" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034624 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b82e1abf-77fe-498c-8a09-2e509a551607" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.034635 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-notification-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034641 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-notification-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.034655 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-central-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034662 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-central-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.034711 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="sg-core" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034719 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="sg-core" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.034726 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="proxy-httpd" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034731 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="proxy-httpd" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034891 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b82e1abf-77fe-498c-8a09-2e509a551607" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034904 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="proxy-httpd" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034915 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-central-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034930 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="sg-core" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.034941 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" containerName="ceilometer-notification-agent" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.035586 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.038270 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.038396 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.038831 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.054591 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.057703 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.062981 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.063207 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.063772 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.068984 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.077774 4837 scope.go:117] "RemoveContainer" containerID="1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.080433 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f\": container with ID starting with 1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f not found: ID does not exist" containerID="1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.080491 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f"} err="failed to get container status \"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f\": rpc error: code = NotFound desc = could not find container \"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f\": container with ID starting with 1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.080519 4837 scope.go:117] "RemoveContainer" containerID="03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.081467 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333\": container with ID starting with 03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333 not found: ID does not exist" containerID="03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.081506 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333"} err="failed to get container status \"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333\": rpc error: code = NotFound desc = could not find container \"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333\": container with ID starting with 03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333 not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.081535 4837 scope.go:117] "RemoveContainer" containerID="43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.081864 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b\": container with ID starting with 43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b not found: ID does not exist" containerID="43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.081889 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b"} err="failed to get container status \"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b\": rpc error: code = NotFound desc = could not find container \"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b\": container with ID starting with 43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.081905 4837 scope.go:117] "RemoveContainer" containerID="a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" Oct 01 07:28:07 crc kubenswrapper[4837]: E1001 07:28:07.082175 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19\": container with ID starting with a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19 not found: ID does not exist" containerID="a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082191 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19"} err="failed to get container status \"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19\": rpc error: code = NotFound desc = could not find container \"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19\": container with ID starting with a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19 not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082202 4837 scope.go:117] "RemoveContainer" containerID="1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082434 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f"} err="failed to get container status \"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f\": rpc error: code = NotFound desc = could not find container \"1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f\": container with ID starting with 1e9399560360db358de9ad533d6ee95ee94b1516d257b29cd675bbe24eb76d4f not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082449 4837 scope.go:117] "RemoveContainer" containerID="03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082635 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333"} err="failed to get container status \"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333\": rpc error: code = NotFound desc = could not find container \"03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333\": container with ID starting with 03893c3fada1d619b811af179570afa7167948bd2f7b711d2139122908704333 not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082650 4837 scope.go:117] "RemoveContainer" containerID="43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082932 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b"} err="failed to get container status \"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b\": rpc error: code = NotFound desc = could not find container \"43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b\": container with ID starting with 43d8cfa80beef8e2c55193eb7e4722d7826575ef87f9a91add1193fc7651e04b not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.082950 4837 scope.go:117] "RemoveContainer" containerID="a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.083233 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19"} err="failed to get container status \"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19\": rpc error: code = NotFound desc = could not find container \"a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19\": container with ID starting with a9bf7291528497136e3f21cee3e0c0e1fb1590ce1b5ad61733ff8ddb36696a19 not found: ID does not exist" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.083590 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124779 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124801 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124818 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtwcf\" (UniqueName: \"kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124897 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.124956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2h2x\" (UniqueName: \"kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125046 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125101 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125117 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125256 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125310 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125347 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.125383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.227805 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.227933 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.227971 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228007 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtwcf\" (UniqueName: \"kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228063 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228107 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2h2x\" (UniqueName: \"kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228224 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228260 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228301 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228336 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228379 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228421 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228624 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.228956 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.234411 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.234427 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.234785 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.235239 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.235353 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.235445 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.237013 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.237918 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.245087 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.248273 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2h2x\" (UniqueName: \"kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x\") pod \"ceilometer-0\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.249415 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtwcf\" (UniqueName: \"kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.371530 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.400337 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.601180 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.833194 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46d312f6-5f1b-4878-ad05-7146f85b77b4" path="/var/lib/kubelet/pods/46d312f6-5f1b-4878-ad05-7146f85b77b4/volumes" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.835797 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b82e1abf-77fe-498c-8a09-2e509a551607" path="/var/lib/kubelet/pods/b82e1abf-77fe-498c-8a09-2e509a551607/volumes" Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.836801 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.905201 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:07 crc kubenswrapper[4837]: W1001 07:28:07.919206 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9af58795_fdb5_432f_be4c_b19c4d1d2c44.slice/crio-c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76 WatchSource:0}: Error finding container c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76: Status 404 returned error can't find the container with id c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76 Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.933379 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddfefe16-d728-4570-9ef6-cc717162f590","Type":"ContainerStarted","Data":"2edc6f515b97e4a41b981ab02aa6490b9169ab572df9b27cfe7a7c6ac4513ba2"} Oct 01 07:28:07 crc kubenswrapper[4837]: I1001 07:28:07.935989 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerStarted","Data":"c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76"} Oct 01 07:28:08 crc kubenswrapper[4837]: I1001 07:28:08.958516 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddfefe16-d728-4570-9ef6-cc717162f590","Type":"ContainerStarted","Data":"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e"} Oct 01 07:28:08 crc kubenswrapper[4837]: I1001 07:28:08.964478 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerStarted","Data":"82f7d563648b7707832c5c7466bc63bb736e1c4e59613467ee367330271cad6d"} Oct 01 07:28:08 crc kubenswrapper[4837]: I1001 07:28:08.987829 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.987813002 podStartE2EDuration="2.987813002s" podCreationTimestamp="2025-10-01 07:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:08.979047795 +0000 UTC m=+1345.820655260" watchObservedRunningTime="2025-10-01 07:28:08.987813002 +0000 UTC m=+1345.829420457" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.410631 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.476053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vtn5\" (UniqueName: \"kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5\") pod \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.476134 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs\") pod \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.476322 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle\") pod \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.476351 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data\") pod \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\" (UID: \"981c4ae4-7e11-4a52-a559-2ebb55c622d5\") " Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.477315 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs" (OuterVolumeSpecName: "logs") pod "981c4ae4-7e11-4a52-a559-2ebb55c622d5" (UID: "981c4ae4-7e11-4a52-a559-2ebb55c622d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.487925 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5" (OuterVolumeSpecName: "kube-api-access-6vtn5") pod "981c4ae4-7e11-4a52-a559-2ebb55c622d5" (UID: "981c4ae4-7e11-4a52-a559-2ebb55c622d5"). InnerVolumeSpecName "kube-api-access-6vtn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.509462 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data" (OuterVolumeSpecName: "config-data") pod "981c4ae4-7e11-4a52-a559-2ebb55c622d5" (UID: "981c4ae4-7e11-4a52-a559-2ebb55c622d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.509992 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "981c4ae4-7e11-4a52-a559-2ebb55c622d5" (UID: "981c4ae4-7e11-4a52-a559-2ebb55c622d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.578592 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/981c4ae4-7e11-4a52-a559-2ebb55c622d5-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.578625 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.578635 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/981c4ae4-7e11-4a52-a559-2ebb55c622d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.578647 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vtn5\" (UniqueName: \"kubernetes.io/projected/981c4ae4-7e11-4a52-a559-2ebb55c622d5-kube-api-access-6vtn5\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.974146 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerStarted","Data":"60a035800ab40246518e36001a9d37d48e3accfcf08a7a2ff8966dd3e863a8e4"} Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.975934 4837 generic.go:334] "Generic (PLEG): container finished" podID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerID="7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a" exitCode=0 Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.976007 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerDied","Data":"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a"} Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.976071 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"981c4ae4-7e11-4a52-a559-2ebb55c622d5","Type":"ContainerDied","Data":"37d31dc9a7e461c65693d223e45838b76ebe19159df9b21662019a09caa515f6"} Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.976100 4837 scope.go:117] "RemoveContainer" containerID="7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.976022 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:09 crc kubenswrapper[4837]: I1001 07:28:09.998028 4837 scope.go:117] "RemoveContainer" containerID="deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.024517 4837 scope.go:117] "RemoveContainer" containerID="7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a" Oct 01 07:28:10 crc kubenswrapper[4837]: E1001 07:28:10.024919 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a\": container with ID starting with 7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a not found: ID does not exist" containerID="7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.024945 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a"} err="failed to get container status \"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a\": rpc error: code = NotFound desc = could not find container \"7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a\": container with ID starting with 7ff6e0f012ea8a8d0417317d008680c37c5cd6938abf8814c727376b4a7e967a not found: ID does not exist" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.024965 4837 scope.go:117] "RemoveContainer" containerID="deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c" Oct 01 07:28:10 crc kubenswrapper[4837]: E1001 07:28:10.025271 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c\": container with ID starting with deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c not found: ID does not exist" containerID="deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.025291 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c"} err="failed to get container status \"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c\": rpc error: code = NotFound desc = could not find container \"deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c\": container with ID starting with deded89db66e528ad0ae1aeef43e79a11dc47012243108751656cce594379d6c not found: ID does not exist" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.032892 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.052298 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.062940 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:10 crc kubenswrapper[4837]: E1001 07:28:10.063354 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-log" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.063367 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-log" Oct 01 07:28:10 crc kubenswrapper[4837]: E1001 07:28:10.063383 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-api" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.063389 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-api" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.063572 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-log" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.063599 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" containerName="nova-api-api" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.071823 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.074110 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.074344 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.078162 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.085709 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114462 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114497 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114553 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114910 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.114943 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgpds\" (UniqueName: \"kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.216561 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.216863 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.217033 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.217182 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.217350 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.217395 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.217462 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgpds\" (UniqueName: \"kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.221323 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.221567 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.222256 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.225113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.238461 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgpds\" (UniqueName: \"kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds\") pod \"nova-api-0\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.390107 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.910245 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.993908 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerStarted","Data":"5ed2be1636cd3ecb4b1b28b8de77bab8aa138fe6f32f2d417e861cbe3e2e5a4a"} Oct 01 07:28:10 crc kubenswrapper[4837]: I1001 07:28:10.996648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerStarted","Data":"be8a214da7bdf20dd67ca2fb56af5ef30d2d16475e5e0495f5801c22e4a0fddb"} Oct 01 07:28:11 crc kubenswrapper[4837]: I1001 07:28:11.827467 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="981c4ae4-7e11-4a52-a559-2ebb55c622d5" path="/var/lib/kubelet/pods/981c4ae4-7e11-4a52-a559-2ebb55c622d5/volumes" Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.005985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerStarted","Data":"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca"} Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.006244 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerStarted","Data":"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567"} Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008372 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerStarted","Data":"e16cb98018d7c03dd0dd1ecd72811712a0e0dd054524e8cbf1d1ad4808a82d5a"} Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008554 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-central-agent" containerID="cri-o://82f7d563648b7707832c5c7466bc63bb736e1c4e59613467ee367330271cad6d" gracePeriod=30 Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008636 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="proxy-httpd" containerID="cri-o://e16cb98018d7c03dd0dd1ecd72811712a0e0dd054524e8cbf1d1ad4808a82d5a" gracePeriod=30 Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008672 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008738 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-notification-agent" containerID="cri-o://60a035800ab40246518e36001a9d37d48e3accfcf08a7a2ff8966dd3e863a8e4" gracePeriod=30 Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.008766 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="sg-core" containerID="cri-o://5ed2be1636cd3ecb4b1b28b8de77bab8aa138fe6f32f2d417e861cbe3e2e5a4a" gracePeriod=30 Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.033230 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.033212399 podStartE2EDuration="2.033212399s" podCreationTimestamp="2025-10-01 07:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:12.025500038 +0000 UTC m=+1348.867107493" watchObservedRunningTime="2025-10-01 07:28:12.033212399 +0000 UTC m=+1348.874819854" Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.058264 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6615746 podStartE2EDuration="6.05824344s" podCreationTimestamp="2025-10-01 07:28:06 +0000 UTC" firstStartedPulling="2025-10-01 07:28:07.923158504 +0000 UTC m=+1344.764765959" lastFinishedPulling="2025-10-01 07:28:11.319827334 +0000 UTC m=+1348.161434799" observedRunningTime="2025-10-01 07:28:12.050362084 +0000 UTC m=+1348.891969549" watchObservedRunningTime="2025-10-01 07:28:12.05824344 +0000 UTC m=+1348.899850905" Oct 01 07:28:12 crc kubenswrapper[4837]: I1001 07:28:12.373091 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028286 4837 generic.go:334] "Generic (PLEG): container finished" podID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerID="e16cb98018d7c03dd0dd1ecd72811712a0e0dd054524e8cbf1d1ad4808a82d5a" exitCode=0 Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028640 4837 generic.go:334] "Generic (PLEG): container finished" podID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerID="5ed2be1636cd3ecb4b1b28b8de77bab8aa138fe6f32f2d417e861cbe3e2e5a4a" exitCode=2 Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028659 4837 generic.go:334] "Generic (PLEG): container finished" podID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerID="60a035800ab40246518e36001a9d37d48e3accfcf08a7a2ff8966dd3e863a8e4" exitCode=0 Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028361 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerDied","Data":"e16cb98018d7c03dd0dd1ecd72811712a0e0dd054524e8cbf1d1ad4808a82d5a"} Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028792 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerDied","Data":"5ed2be1636cd3ecb4b1b28b8de77bab8aa138fe6f32f2d417e861cbe3e2e5a4a"} Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.028823 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerDied","Data":"60a035800ab40246518e36001a9d37d48e3accfcf08a7a2ff8966dd3e863a8e4"} Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.431938 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.514772 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:28:13 crc kubenswrapper[4837]: I1001 07:28:13.515824 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="dnsmasq-dns" containerID="cri-o://f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08" gracePeriod=10 Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.037214 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.039844 4837 generic.go:334] "Generic (PLEG): container finished" podID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerID="f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08" exitCode=0 Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.039932 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" event={"ID":"1c7351cc-3937-4008-a199-5fe7ccaa6276","Type":"ContainerDied","Data":"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08"} Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.039959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" event={"ID":"1c7351cc-3937-4008-a199-5fe7ccaa6276","Type":"ContainerDied","Data":"1ff26642dab94cc1b8980302658e5c2f6625d83998967aaeb0abfc38d72f6ce5"} Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.039976 4837 scope.go:117] "RemoveContainer" containerID="f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.086579 4837 scope.go:117] "RemoveContainer" containerID="7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098225 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098283 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098541 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlvsp\" (UniqueName: \"kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098641 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098741 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.098796 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config\") pod \"1c7351cc-3937-4008-a199-5fe7ccaa6276\" (UID: \"1c7351cc-3937-4008-a199-5fe7ccaa6276\") " Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.104516 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp" (OuterVolumeSpecName: "kube-api-access-tlvsp") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "kube-api-access-tlvsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.136995 4837 scope.go:117] "RemoveContainer" containerID="f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08" Oct 01 07:28:14 crc kubenswrapper[4837]: E1001 07:28:14.145841 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08\": container with ID starting with f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08 not found: ID does not exist" containerID="f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.145877 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08"} err="failed to get container status \"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08\": rpc error: code = NotFound desc = could not find container \"f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08\": container with ID starting with f69ad60bcdb7d11d6302e90a97212d652b93e847e479a0c7ba5873ce42996a08 not found: ID does not exist" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.145901 4837 scope.go:117] "RemoveContainer" containerID="7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57" Oct 01 07:28:14 crc kubenswrapper[4837]: E1001 07:28:14.147437 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57\": container with ID starting with 7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57 not found: ID does not exist" containerID="7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.147485 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57"} err="failed to get container status \"7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57\": rpc error: code = NotFound desc = could not find container \"7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57\": container with ID starting with 7ac2ba78a58387861f24ab8d47e3bd6cbf8e5e1fe5a8b21456b4c4e701498f57 not found: ID does not exist" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.161674 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.168680 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config" (OuterVolumeSpecName: "config") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.169653 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.179412 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.187257 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c7351cc-3937-4008-a199-5fe7ccaa6276" (UID: "1c7351cc-3937-4008-a199-5fe7ccaa6276"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205012 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlvsp\" (UniqueName: \"kubernetes.io/projected/1c7351cc-3937-4008-a199-5fe7ccaa6276-kube-api-access-tlvsp\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205043 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205053 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205062 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205070 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:14 crc kubenswrapper[4837]: I1001 07:28:14.205078 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c7351cc-3937-4008-a199-5fe7ccaa6276-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.057454 4837 generic.go:334] "Generic (PLEG): container finished" podID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerID="82f7d563648b7707832c5c7466bc63bb736e1c4e59613467ee367330271cad6d" exitCode=0 Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.057614 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerDied","Data":"82f7d563648b7707832c5c7466bc63bb736e1c4e59613467ee367330271cad6d"} Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.057791 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9af58795-fdb5-432f-be4c-b19c4d1d2c44","Type":"ContainerDied","Data":"c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76"} Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.057830 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c164b6d25a5abec95d43ff3d5f92a6d39efba7d6d39a5aacd163421d9025dd76" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.058682 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f5d75d8c-7z2cd" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.072454 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.088058 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.094625 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f5d75d8c-7z2cd"] Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.123874 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.123956 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.123985 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.124048 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.124074 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2h2x\" (UniqueName: \"kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.124104 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.124136 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.124216 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data\") pod \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\" (UID: \"9af58795-fdb5-432f-be4c-b19c4d1d2c44\") " Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.125455 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.125917 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.132322 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x" (OuterVolumeSpecName: "kube-api-access-v2h2x") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "kube-api-access-v2h2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.146781 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts" (OuterVolumeSpecName: "scripts") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.173914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.177212 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.198492 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226515 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226548 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226558 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226566 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226575 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2h2x\" (UniqueName: \"kubernetes.io/projected/9af58795-fdb5-432f-be4c-b19c4d1d2c44-kube-api-access-v2h2x\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226584 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9af58795-fdb5-432f-be4c-b19c4d1d2c44-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.226593 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.243251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data" (OuterVolumeSpecName: "config-data") pod "9af58795-fdb5-432f-be4c-b19c4d1d2c44" (UID: "9af58795-fdb5-432f-be4c-b19c4d1d2c44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.328300 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9af58795-fdb5-432f-be4c-b19c4d1d2c44-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:15 crc kubenswrapper[4837]: I1001 07:28:15.836447 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" path="/var/lib/kubelet/pods/1c7351cc-3937-4008-a199-5fe7ccaa6276/volumes" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.067893 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.097454 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.105990 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.137918 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138395 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="init" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138416 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="init" Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138434 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-central-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138443 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-central-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138463 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="proxy-httpd" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138474 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="proxy-httpd" Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138495 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="sg-core" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138503 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="sg-core" Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138522 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-notification-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138531 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-notification-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: E1001 07:28:16.138549 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="dnsmasq-dns" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138557 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="dnsmasq-dns" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138782 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="proxy-httpd" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138804 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-notification-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138817 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7351cc-3937-4008-a199-5fe7ccaa6276" containerName="dnsmasq-dns" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138832 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="ceilometer-central-agent" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.138853 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" containerName="sg-core" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.140875 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.144920 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.145436 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.146001 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.160908 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.246101 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.246466 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.246667 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.246894 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.247001 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.247670 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.247821 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqzlh\" (UniqueName: \"kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.247866 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.349820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.349933 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.349983 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.350056 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.350097 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqzlh\" (UniqueName: \"kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.350129 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.350184 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.350303 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.351002 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.351027 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.355360 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.355408 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.355595 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.355924 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.362140 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.369289 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqzlh\" (UniqueName: \"kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh\") pod \"ceilometer-0\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.466000 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:28:16 crc kubenswrapper[4837]: W1001 07:28:16.987958 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17575b0a_a27e_4cc5_a84a_0ed9adf0ed65.slice/crio-8414978f862c1ae0fe709e1d354a9339ff84e7723f35ffc1823a5b043b3ae0bd WatchSource:0}: Error finding container 8414978f862c1ae0fe709e1d354a9339ff84e7723f35ffc1823a5b043b3ae0bd: Status 404 returned error can't find the container with id 8414978f862c1ae0fe709e1d354a9339ff84e7723f35ffc1823a5b043b3ae0bd Oct 01 07:28:16 crc kubenswrapper[4837]: I1001 07:28:16.989576 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:28:17 crc kubenswrapper[4837]: I1001 07:28:17.080589 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerStarted","Data":"8414978f862c1ae0fe709e1d354a9339ff84e7723f35ffc1823a5b043b3ae0bd"} Oct 01 07:28:17 crc kubenswrapper[4837]: I1001 07:28:17.374049 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:17 crc kubenswrapper[4837]: I1001 07:28:17.398480 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:17 crc kubenswrapper[4837]: I1001 07:28:17.835493 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9af58795-fdb5-432f-be4c-b19c4d1d2c44" path="/var/lib/kubelet/pods/9af58795-fdb5-432f-be4c-b19c4d1d2c44/volumes" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.094002 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerStarted","Data":"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db"} Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.118304 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.298743 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bphz9"] Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.302033 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.304160 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.304320 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.321835 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bphz9"] Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.393720 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp22w\" (UniqueName: \"kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.393809 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.393841 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.393857 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.495343 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp22w\" (UniqueName: \"kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.495454 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.495487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.495505 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.501664 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.502247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.507803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.514046 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp22w\" (UniqueName: \"kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w\") pod \"nova-cell1-cell-mapping-bphz9\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:18 crc kubenswrapper[4837]: I1001 07:28:18.616494 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:19 crc kubenswrapper[4837]: I1001 07:28:19.090309 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bphz9"] Oct 01 07:28:19 crc kubenswrapper[4837]: W1001 07:28:19.096990 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08d5a60a_28ff_4ce9_858e_e6835a77b65f.slice/crio-bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0 WatchSource:0}: Error finding container bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0: Status 404 returned error can't find the container with id bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0 Oct 01 07:28:19 crc kubenswrapper[4837]: I1001 07:28:19.107835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerStarted","Data":"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40"} Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.121405 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerStarted","Data":"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a"} Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.123724 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bphz9" event={"ID":"08d5a60a-28ff-4ce9-858e-e6835a77b65f","Type":"ContainerStarted","Data":"ed6ba9d9762a04643429cdf39156487cb7939eb11057b92327a090e24655643f"} Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.123786 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bphz9" event={"ID":"08d5a60a-28ff-4ce9-858e-e6835a77b65f","Type":"ContainerStarted","Data":"bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0"} Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.154083 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bphz9" podStartSLOduration=2.154058525 podStartE2EDuration="2.154058525s" podCreationTimestamp="2025-10-01 07:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:20.142849787 +0000 UTC m=+1356.984457262" watchObservedRunningTime="2025-10-01 07:28:20.154058525 +0000 UTC m=+1356.995665990" Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.403193 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:28:20 crc kubenswrapper[4837]: I1001 07:28:20.403248 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:28:21 crc kubenswrapper[4837]: I1001 07:28:21.139885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerStarted","Data":"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c"} Oct 01 07:28:21 crc kubenswrapper[4837]: I1001 07:28:21.174392 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.454065506 podStartE2EDuration="5.174368143s" podCreationTimestamp="2025-10-01 07:28:16 +0000 UTC" firstStartedPulling="2025-10-01 07:28:16.995913239 +0000 UTC m=+1353.837520704" lastFinishedPulling="2025-10-01 07:28:20.716215886 +0000 UTC m=+1357.557823341" observedRunningTime="2025-10-01 07:28:21.173340947 +0000 UTC m=+1358.014948412" watchObservedRunningTime="2025-10-01 07:28:21.174368143 +0000 UTC m=+1358.015975628" Oct 01 07:28:21 crc kubenswrapper[4837]: I1001 07:28:21.417967 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:21 crc kubenswrapper[4837]: I1001 07:28:21.417978 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:22 crc kubenswrapper[4837]: I1001 07:28:22.155629 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 07:28:24 crc kubenswrapper[4837]: I1001 07:28:24.187575 4837 generic.go:334] "Generic (PLEG): container finished" podID="08d5a60a-28ff-4ce9-858e-e6835a77b65f" containerID="ed6ba9d9762a04643429cdf39156487cb7939eb11057b92327a090e24655643f" exitCode=0 Oct 01 07:28:24 crc kubenswrapper[4837]: I1001 07:28:24.187752 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bphz9" event={"ID":"08d5a60a-28ff-4ce9-858e-e6835a77b65f","Type":"ContainerDied","Data":"ed6ba9d9762a04643429cdf39156487cb7939eb11057b92327a090e24655643f"} Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.734611 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.797533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp22w\" (UniqueName: \"kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w\") pod \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.797734 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data\") pod \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.797776 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts\") pod \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.797805 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle\") pod \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\" (UID: \"08d5a60a-28ff-4ce9-858e-e6835a77b65f\") " Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.803029 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w" (OuterVolumeSpecName: "kube-api-access-hp22w") pod "08d5a60a-28ff-4ce9-858e-e6835a77b65f" (UID: "08d5a60a-28ff-4ce9-858e-e6835a77b65f"). InnerVolumeSpecName "kube-api-access-hp22w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.803717 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts" (OuterVolumeSpecName: "scripts") pod "08d5a60a-28ff-4ce9-858e-e6835a77b65f" (UID: "08d5a60a-28ff-4ce9-858e-e6835a77b65f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.824091 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data" (OuterVolumeSpecName: "config-data") pod "08d5a60a-28ff-4ce9-858e-e6835a77b65f" (UID: "08d5a60a-28ff-4ce9-858e-e6835a77b65f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.847407 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08d5a60a-28ff-4ce9-858e-e6835a77b65f" (UID: "08d5a60a-28ff-4ce9-858e-e6835a77b65f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.900518 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.900713 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.900789 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d5a60a-28ff-4ce9-858e-e6835a77b65f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:25 crc kubenswrapper[4837]: I1001 07:28:25.900856 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp22w\" (UniqueName: \"kubernetes.io/projected/08d5a60a-28ff-4ce9-858e-e6835a77b65f-kube-api-access-hp22w\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.216383 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bphz9" event={"ID":"08d5a60a-28ff-4ce9-858e-e6835a77b65f","Type":"ContainerDied","Data":"bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0"} Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.216437 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bca3ecdc533fe8a895fa5ee2a1daebab203b3f68902873c5fa8e7dc998d8e9a0" Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.216472 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bphz9" Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.420500 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.420895 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-log" containerID="cri-o://60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567" gracePeriod=30 Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.421050 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-api" containerID="cri-o://fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca" gracePeriod=30 Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.504263 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.504625 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" containerID="cri-o://8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13" gracePeriod=30 Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.504732 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" containerID="cri-o://17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea" gracePeriod=30 Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.518639 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:26 crc kubenswrapper[4837]: I1001 07:28:26.518863 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerName="nova-scheduler-scheduler" containerID="cri-o://589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" gracePeriod=30 Oct 01 07:28:27 crc kubenswrapper[4837]: I1001 07:28:27.231252 4837 generic.go:334] "Generic (PLEG): container finished" podID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerID="8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13" exitCode=143 Oct 01 07:28:27 crc kubenswrapper[4837]: I1001 07:28:27.231294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerDied","Data":"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13"} Oct 01 07:28:27 crc kubenswrapper[4837]: I1001 07:28:27.234489 4837 generic.go:334] "Generic (PLEG): container finished" podID="8cf5beff-416d-481f-804f-8feb954840e3" containerID="60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567" exitCode=143 Oct 01 07:28:27 crc kubenswrapper[4837]: I1001 07:28:27.234534 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerDied","Data":"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567"} Oct 01 07:28:28 crc kubenswrapper[4837]: E1001 07:28:28.019303 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:28:28 crc kubenswrapper[4837]: E1001 07:28:28.022499 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:28:28 crc kubenswrapper[4837]: E1001 07:28:28.024586 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:28:28 crc kubenswrapper[4837]: E1001 07:28:28.024658 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerName="nova-scheduler-scheduler" Oct 01 07:28:29 crc kubenswrapper[4837]: I1001 07:28:29.641308 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:33868->10.217.0.190:8775: read: connection reset by peer" Oct 01 07:28:29 crc kubenswrapper[4837]: I1001 07:28:29.641395 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:33878->10.217.0.190:8775: read: connection reset by peer" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.103170 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.151596 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.189928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190065 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190098 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190316 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190374 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190403 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgpds\" (UniqueName: \"kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds\") pod \"8cf5beff-416d-481f-804f-8feb954840e3\" (UID: \"8cf5beff-416d-481f-804f-8feb954840e3\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.190786 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs" (OuterVolumeSpecName: "logs") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.191002 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cf5beff-416d-481f-804f-8feb954840e3-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.199942 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds" (OuterVolumeSpecName: "kube-api-access-mgpds") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "kube-api-access-mgpds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.224946 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data" (OuterVolumeSpecName: "config-data") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.228998 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.248775 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.271518 4837 generic.go:334] "Generic (PLEG): container finished" podID="8cf5beff-416d-481f-804f-8feb954840e3" containerID="fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca" exitCode=0 Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.271573 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerDied","Data":"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca"} Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.271600 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8cf5beff-416d-481f-804f-8feb954840e3","Type":"ContainerDied","Data":"be8a214da7bdf20dd67ca2fb56af5ef30d2d16475e5e0495f5801c22e4a0fddb"} Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.271615 4837 scope.go:117] "RemoveContainer" containerID="fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.271747 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.273061 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8cf5beff-416d-481f-804f-8feb954840e3" (UID: "8cf5beff-416d-481f-804f-8feb954840e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.274949 4837 generic.go:334] "Generic (PLEG): container finished" podID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerID="17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea" exitCode=0 Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.274988 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.274995 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerDied","Data":"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea"} Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.275103 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e3e952ac-c2ad-4b22-9eab-19c5f36f703c","Type":"ContainerDied","Data":"a697c9abc111e268e4beba0c4f657072bde1cba810dff254aac8c36f92e79835"} Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.292571 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs\") pod \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.292638 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data\") pod \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.292842 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle\") pod \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.292894 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs\") pod \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.292966 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8p69\" (UniqueName: \"kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69\") pod \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\" (UID: \"e3e952ac-c2ad-4b22-9eab-19c5f36f703c\") " Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293322 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293339 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293349 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgpds\" (UniqueName: \"kubernetes.io/projected/8cf5beff-416d-481f-804f-8feb954840e3-kube-api-access-mgpds\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293360 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293368 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cf5beff-416d-481f-804f-8feb954840e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.293947 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs" (OuterVolumeSpecName: "logs") pod "e3e952ac-c2ad-4b22-9eab-19c5f36f703c" (UID: "e3e952ac-c2ad-4b22-9eab-19c5f36f703c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.296827 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69" (OuterVolumeSpecName: "kube-api-access-z8p69") pod "e3e952ac-c2ad-4b22-9eab-19c5f36f703c" (UID: "e3e952ac-c2ad-4b22-9eab-19c5f36f703c"). InnerVolumeSpecName "kube-api-access-z8p69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.301724 4837 scope.go:117] "RemoveContainer" containerID="60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.319383 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3e952ac-c2ad-4b22-9eab-19c5f36f703c" (UID: "e3e952ac-c2ad-4b22-9eab-19c5f36f703c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.326302 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data" (OuterVolumeSpecName: "config-data") pod "e3e952ac-c2ad-4b22-9eab-19c5f36f703c" (UID: "e3e952ac-c2ad-4b22-9eab-19c5f36f703c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.330480 4837 scope.go:117] "RemoveContainer" containerID="fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.330974 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca\": container with ID starting with fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca not found: ID does not exist" containerID="fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.331011 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca"} err="failed to get container status \"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca\": rpc error: code = NotFound desc = could not find container \"fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca\": container with ID starting with fdd94cd3260f9fa3aca9cd8d6cff5c655fd9cd7dbfa2e50cc082ee7eacc01fca not found: ID does not exist" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.331035 4837 scope.go:117] "RemoveContainer" containerID="60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.331407 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567\": container with ID starting with 60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567 not found: ID does not exist" containerID="60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.331441 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567"} err="failed to get container status \"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567\": rpc error: code = NotFound desc = could not find container \"60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567\": container with ID starting with 60e3f89be16866c197c2f9c2ea73c8b400d68fd1078c49cb5bf2cbddf5a58567 not found: ID does not exist" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.331463 4837 scope.go:117] "RemoveContainer" containerID="17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.349762 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e3e952ac-c2ad-4b22-9eab-19c5f36f703c" (UID: "e3e952ac-c2ad-4b22-9eab-19c5f36f703c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.357383 4837 scope.go:117] "RemoveContainer" containerID="8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.378749 4837 scope.go:117] "RemoveContainer" containerID="17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.379172 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea\": container with ID starting with 17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea not found: ID does not exist" containerID="17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.379201 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea"} err="failed to get container status \"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea\": rpc error: code = NotFound desc = could not find container \"17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea\": container with ID starting with 17357d02e8e4b34f81ed327e39963388bbacf2d166bdde2f2a52f46d95bc62ea not found: ID does not exist" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.379220 4837 scope.go:117] "RemoveContainer" containerID="8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.379431 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13\": container with ID starting with 8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13 not found: ID does not exist" containerID="8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.379445 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13"} err="failed to get container status \"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13\": rpc error: code = NotFound desc = could not find container \"8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13\": container with ID starting with 8cd2326f9248260d808ca9328ebc5d7866534b581addddcd6fe43e8335c1ac13 not found: ID does not exist" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.394585 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.394603 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8p69\" (UniqueName: \"kubernetes.io/projected/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-kube-api-access-z8p69\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.394613 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.394622 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.394632 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e952ac-c2ad-4b22-9eab-19c5f36f703c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.621807 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.634989 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.645836 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.657772 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665256 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.665756 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-log" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665776 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-log" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.665794 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665803 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.665822 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665832 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.665876 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-api" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665885 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-api" Oct 01 07:28:30 crc kubenswrapper[4837]: E1001 07:28:30.665907 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d5a60a-28ff-4ce9-858e-e6835a77b65f" containerName="nova-manage" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.665916 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d5a60a-28ff-4ce9-858e-e6835a77b65f" containerName="nova-manage" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.666141 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-api" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.666166 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-metadata" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.666191 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" containerName="nova-metadata-log" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.666210 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf5beff-416d-481f-804f-8feb954840e3" containerName="nova-api-log" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.666224 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d5a60a-28ff-4ce9-858e-e6835a77b65f" containerName="nova-manage" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.667484 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.670474 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.670683 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.672066 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.673611 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.676280 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.676512 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.676612 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.679405 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.714003 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805281 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805342 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805365 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzrnb\" (UniqueName: \"kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805555 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805577 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdj6t\" (UniqueName: \"kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805627 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805646 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.805719 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907617 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907711 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907736 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907764 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907819 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzrnb\" (UniqueName: \"kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907898 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907962 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.907998 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdj6t\" (UniqueName: \"kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.908032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.908057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.908158 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.909838 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.910793 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.915284 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.917456 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.918841 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.919169 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.920417 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.931095 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.934187 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.943509 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdj6t\" (UniqueName: \"kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t\") pod \"nova-api-0\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " pod="openstack/nova-api-0" Oct 01 07:28:30 crc kubenswrapper[4837]: I1001 07:28:30.943851 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzrnb\" (UniqueName: \"kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb\") pod \"nova-metadata-0\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " pod="openstack/nova-metadata-0" Oct 01 07:28:31 crc kubenswrapper[4837]: I1001 07:28:31.061940 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:28:31 crc kubenswrapper[4837]: I1001 07:28:31.066914 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:28:31 crc kubenswrapper[4837]: I1001 07:28:31.308585 4837 generic.go:334] "Generic (PLEG): container finished" podID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerID="589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" exitCode=0 Oct 01 07:28:31 crc kubenswrapper[4837]: I1001 07:28:31.308790 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83caf5e4-336c-4def-9eb6-ba7ad607cfb2","Type":"ContainerDied","Data":"589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611"} Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.414391 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.524962 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68758\" (UniqueName: \"kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758\") pod \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.525357 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data\") pod \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.525383 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle\") pod \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\" (UID: \"83caf5e4-336c-4def-9eb6-ba7ad607cfb2\") " Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.533134 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758" (OuterVolumeSpecName: "kube-api-access-68758") pod "83caf5e4-336c-4def-9eb6-ba7ad607cfb2" (UID: "83caf5e4-336c-4def-9eb6-ba7ad607cfb2"). InnerVolumeSpecName "kube-api-access-68758". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.552839 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data" (OuterVolumeSpecName: "config-data") pod "83caf5e4-336c-4def-9eb6-ba7ad607cfb2" (UID: "83caf5e4-336c-4def-9eb6-ba7ad607cfb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.555211 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83caf5e4-336c-4def-9eb6-ba7ad607cfb2" (UID: "83caf5e4-336c-4def-9eb6-ba7ad607cfb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.627836 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.627868 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.627883 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68758\" (UniqueName: \"kubernetes.io/projected/83caf5e4-336c-4def-9eb6-ba7ad607cfb2-kube-api-access-68758\") on node \"crc\" DevicePath \"\"" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.830815 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cf5beff-416d-481f-804f-8feb954840e3" path="/var/lib/kubelet/pods/8cf5beff-416d-481f-804f-8feb954840e3/volumes" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:31.832140 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e952ac-c2ad-4b22-9eab-19c5f36f703c" path="/var/lib/kubelet/pods/e3e952ac-c2ad-4b22-9eab-19c5f36f703c/volumes" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.146995 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.168629 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: W1001 07:28:32.177554 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08552fbf_b050_46ec_b8bf_83cde04f87e6.slice/crio-c420ce38f93c3996d8ce993dda04baa9ba1172bb04b4072544170d6e860a8507 WatchSource:0}: Error finding container c420ce38f93c3996d8ce993dda04baa9ba1172bb04b4072544170d6e860a8507: Status 404 returned error can't find the container with id c420ce38f93c3996d8ce993dda04baa9ba1172bb04b4072544170d6e860a8507 Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.326531 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerStarted","Data":"c420ce38f93c3996d8ce993dda04baa9ba1172bb04b4072544170d6e860a8507"} Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.329812 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83caf5e4-336c-4def-9eb6-ba7ad607cfb2","Type":"ContainerDied","Data":"27aad9ded82ee9c5febd8f30d69e6025ee77e885614a14d7fc6112981b263a70"} Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.329855 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.329887 4837 scope.go:117] "RemoveContainer" containerID="589ed6d1ec3d30774464579f1359d51f4a5151d904def17a37a70ca5669b7611" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.336988 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerStarted","Data":"2c968f6fbc476f7bf90bae775aa43cc0ec58edfa101b4488fdb46bf73a31717b"} Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.380857 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.394999 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.406943 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: E1001 07:28:32.407563 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerName="nova-scheduler-scheduler" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.407587 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerName="nova-scheduler-scheduler" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.407935 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" containerName="nova-scheduler-scheduler" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.409019 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.411992 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.414771 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.543235 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.543393 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.543445 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bx9\" (UniqueName: \"kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.645075 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.645170 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.645205 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bx9\" (UniqueName: \"kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.651445 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.652149 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.663304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bx9\" (UniqueName: \"kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9\") pod \"nova-scheduler-0\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " pod="openstack/nova-scheduler-0" Oct 01 07:28:32 crc kubenswrapper[4837]: I1001 07:28:32.741277 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.200799 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:28:33 crc kubenswrapper[4837]: W1001 07:28:33.210997 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f15a72f_f4ec_4e07_ae6d_7eef5700f543.slice/crio-82ebaffe8662e75b736bd799fabae478db52300184689fa90da793dc63211dd4 WatchSource:0}: Error finding container 82ebaffe8662e75b736bd799fabae478db52300184689fa90da793dc63211dd4: Status 404 returned error can't find the container with id 82ebaffe8662e75b736bd799fabae478db52300184689fa90da793dc63211dd4 Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.349505 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerStarted","Data":"e4a74db66f683da3b67436f59d8f19271ded325f5539b6a7b61f313bce201a1c"} Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.349559 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerStarted","Data":"2422f422ed0b438a27d2a54ccb39f04357a6213eac4f80a06db39e4b2ebad2be"} Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.355219 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerStarted","Data":"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191"} Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.355250 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerStarted","Data":"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a"} Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.357490 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f15a72f-f4ec-4e07-ae6d-7eef5700f543","Type":"ContainerStarted","Data":"82ebaffe8662e75b736bd799fabae478db52300184689fa90da793dc63211dd4"} Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.380324 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.380306626 podStartE2EDuration="3.380306626s" podCreationTimestamp="2025-10-01 07:28:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:33.373750383 +0000 UTC m=+1370.215357848" watchObservedRunningTime="2025-10-01 07:28:33.380306626 +0000 UTC m=+1370.221914081" Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.406065 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.406046945 podStartE2EDuration="3.406046945s" podCreationTimestamp="2025-10-01 07:28:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:33.403755988 +0000 UTC m=+1370.245363453" watchObservedRunningTime="2025-10-01 07:28:33.406046945 +0000 UTC m=+1370.247654400" Oct 01 07:28:33 crc kubenswrapper[4837]: I1001 07:28:33.836553 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83caf5e4-336c-4def-9eb6-ba7ad607cfb2" path="/var/lib/kubelet/pods/83caf5e4-336c-4def-9eb6-ba7ad607cfb2/volumes" Oct 01 07:28:34 crc kubenswrapper[4837]: I1001 07:28:34.374336 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f15a72f-f4ec-4e07-ae6d-7eef5700f543","Type":"ContainerStarted","Data":"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088"} Oct 01 07:28:34 crc kubenswrapper[4837]: I1001 07:28:34.433285 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.433254925 podStartE2EDuration="2.433254925s" podCreationTimestamp="2025-10-01 07:28:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 07:28:34.407088384 +0000 UTC m=+1371.248695849" watchObservedRunningTime="2025-10-01 07:28:34.433254925 +0000 UTC m=+1371.274862410" Oct 01 07:28:36 crc kubenswrapper[4837]: I1001 07:28:36.062613 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 07:28:36 crc kubenswrapper[4837]: I1001 07:28:36.063047 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 07:28:37 crc kubenswrapper[4837]: I1001 07:28:37.741586 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 07:28:41 crc kubenswrapper[4837]: I1001 07:28:41.062366 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 07:28:41 crc kubenswrapper[4837]: I1001 07:28:41.062447 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 07:28:41 crc kubenswrapper[4837]: I1001 07:28:41.067966 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:28:41 crc kubenswrapper[4837]: I1001 07:28:41.068033 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.082872 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.097076 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.097195 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.097104 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.742283 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 07:28:42 crc kubenswrapper[4837]: I1001 07:28:42.778154 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 07:28:43 crc kubenswrapper[4837]: I1001 07:28:43.513359 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 07:28:46 crc kubenswrapper[4837]: I1001 07:28:46.482510 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.072391 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.076865 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.085732 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.086257 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.087646 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.090646 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.098065 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.579531 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.583895 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 07:28:51 crc kubenswrapper[4837]: I1001 07:28:51.589962 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.011780 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.013306 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" containerName="openstackclient" containerID="cri-o://c5da1cb2cb85875baa579d1f9fd8e420c616437d114859950730be4875ac535e" gracePeriod=2 Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.038875 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.282384 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.342206 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.342573 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" containerName="openstackclient" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.342587 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" containerName="openstackclient" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.342809 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" containerName="openstackclient" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.343367 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.379597 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.380323 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="openstack-network-exporter" containerID="cri-o://167c29409de05ee8281b1aed4f4b8f38acfffec5ceb180aebcf6ac61a6a016c5" gracePeriod=300 Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.391959 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.392007 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data podName:96e5feff-74ce-434a-8606-85c51844deb9 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:11.891995048 +0000 UTC m=+1408.733602503 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data") pod "rabbitmq-cell1-server-0" (UID: "96e5feff-74ce-434a-8606-85c51844deb9") : configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.406626 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.436116 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.437306 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.448136 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.493313 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-689ks\" (UniqueName: \"kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks\") pod \"placement7d99-account-delete-pnxgc\" (UID: \"dce4369c-e009-4ae1-bc07-65dff0005807\") " pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.514541 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.516205 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="ovsdbserver-sb" containerID="cri-o://48c9eecb63371cb6c6437405cb59dd74a49403ac0f542fdf6ce9200b6b395b32" gracePeriod=300 Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.517138 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.537834 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.561190 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.597795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-689ks\" (UniqueName: \"kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks\") pod \"placement7d99-account-delete-pnxgc\" (UID: \"dce4369c-e009-4ae1-bc07-65dff0005807\") " pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.597859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlnwg\" (UniqueName: \"kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg\") pod \"cindere1a9-account-delete-qw9xn\" (UID: \"8f72fdf5-3656-49e2-80a8-37d0d65ec212\") " pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.664436 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-689ks\" (UniqueName: \"kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks\") pod \"placement7d99-account-delete-pnxgc\" (UID: \"dce4369c-e009-4ae1-bc07-65dff0005807\") " pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.664836 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.700663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.700737 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pxbv\" (UniqueName: \"kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.700813 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.700868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlnwg\" (UniqueName: \"kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg\") pod \"cindere1a9-account-delete-qw9xn\" (UID: \"8f72fdf5-3656-49e2-80a8-37d0d65ec212\") " pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.701975 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.702019 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data podName:0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:12.202006586 +0000 UTC m=+1409.043614041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data") pod "rabbitmq-server-0" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087") : configmap "rabbitmq-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.766349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlnwg\" (UniqueName: \"kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg\") pod \"cindere1a9-account-delete-qw9xn\" (UID: \"8f72fdf5-3656-49e2-80a8-37d0d65ec212\") " pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.769209 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.804968 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.805014 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pxbv\" (UniqueName: \"kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.805063 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.805654 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.806006 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.811929 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.813078 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.851745 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pxbv\" (UniqueName: \"kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv\") pod \"redhat-operators-xg94g\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.864794 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.864832 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5q9zm"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.887640 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.888192 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8fqtj"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.908423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qkl7\" (UniqueName: \"kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7\") pod \"neutron8fcf-account-delete-g5pcd\" (UID: \"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8\") " pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.910346 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: E1001 07:29:11.910395 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data podName:96e5feff-74ce-434a-8606-85c51844deb9 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:12.910379651 +0000 UTC m=+1409.751987106 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data") pod "rabbitmq-cell1-server-0" (UID: "96e5feff-74ce-434a-8606-85c51844deb9") : configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.919433 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5q9zm"] Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.947924 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_144c538f-4172-4043-b716-62f91b64f8fc/ovsdbserver-sb/0.log" Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.947990 4837 generic.go:334] "Generic (PLEG): container finished" podID="144c538f-4172-4043-b716-62f91b64f8fc" containerID="167c29409de05ee8281b1aed4f4b8f38acfffec5ceb180aebcf6ac61a6a016c5" exitCode=2 Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.948017 4837 generic.go:334] "Generic (PLEG): container finished" podID="144c538f-4172-4043-b716-62f91b64f8fc" containerID="48c9eecb63371cb6c6437405cb59dd74a49403ac0f542fdf6ce9200b6b395b32" exitCode=143 Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.948043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerDied","Data":"167c29409de05ee8281b1aed4f4b8f38acfffec5ceb180aebcf6ac61a6a016c5"} Oct 01 07:29:11 crc kubenswrapper[4837]: I1001 07:29:11.948073 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerDied","Data":"48c9eecb63371cb6c6437405cb59dd74a49403ac0f542fdf6ce9200b6b395b32"} Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.040529 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qkl7\" (UniqueName: \"kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7\") pod \"neutron8fcf-account-delete-g5pcd\" (UID: \"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8\") " pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.089015 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qkl7\" (UniqueName: \"kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7\") pod \"neutron8fcf-account-delete-g5pcd\" (UID: \"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8\") " pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.089113 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8fqtj"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.111874 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.112114 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" containerID="cri-o://648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" gracePeriod=30 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.112536 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="openstack-network-exporter" containerID="cri-o://a699f84318a8fa1ea08dd47672464162bcd065e0a5b23d594adac792bb355f6b" gracePeriod=30 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.192319 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.193935 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.225758 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.226218 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="openstack-network-exporter" containerID="cri-o://a33be5ffcfa7d1c7be08f25ec05a88d909c111fd9dde1c9328ba84acd170fc4c" gracePeriod=300 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.251383 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.256918 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.264734 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t446f\" (UniqueName: \"kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f\") pod \"glance6315-account-delete-ddhmt\" (UID: \"595b1839-8525-4b91-ab30-fd800710cf31\") " pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.265098 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.265141 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data podName:0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:13.265127031 +0000 UTC m=+1410.106734486 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data") pod "rabbitmq-server-0" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087") : configmap "rabbitmq-config-data" not found Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.299763 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.300020 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-2vw77" podUID="28035a28-28d4-46b3-b269-97a523cf731e" containerName="openstack-network-exporter" containerID="cri-o://fa33ac993198b8bd4e6712d1dfc94ce0a3b46c83683d716c97d0c41883173646" gracePeriod=30 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.335454 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.354357 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.369820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t446f\" (UniqueName: \"kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f\") pod \"glance6315-account-delete-ddhmt\" (UID: \"595b1839-8525-4b91-ab30-fd800710cf31\") " pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.384052 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.385753 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.405183 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.413105 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.417549 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:12 crc kubenswrapper[4837]: E1001 07:29:12.417828 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.444614 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t446f\" (UniqueName: \"kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f\") pod \"glance6315-account-delete-ddhmt\" (UID: \"595b1839-8525-4b91-ab30-fd800710cf31\") " pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.451796 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.474678 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkkt7\" (UniqueName: \"kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7\") pod \"barbicanf4b1-account-delete-x448d\" (UID: \"aa5e54c4-ad98-4f26-9681-189997f3bd64\") " pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.493546 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="ovsdbserver-nb" containerID="cri-o://b2dd01f2720f0c29fc64551b59dc951b915dcc75366896c51b0ec3db97f15550" gracePeriod=300 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.508811 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-cx8fl"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.557956 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-cx8fl"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.583891 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkkt7\" (UniqueName: \"kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7\") pod \"barbicanf4b1-account-delete-x448d\" (UID: \"aa5e54c4-ad98-4f26-9681-189997f3bd64\") " pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.603420 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.632752 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.634383 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.651843 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkkt7\" (UniqueName: \"kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7\") pod \"barbicanf4b1-account-delete-x448d\" (UID: \"aa5e54c4-ad98-4f26-9681-189997f3bd64\") " pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.658779 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.686767 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkz9s\" (UniqueName: \"kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s\") pod \"novaapic192-account-delete-qzdmf\" (UID: \"6873cb35-815f-44ae-b38d-bb7ceb806b9d\") " pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.706058 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.726212 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ksdnf"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.776872 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ksdnf"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.797432 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkz9s\" (UniqueName: \"kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s\") pod \"novaapic192-account-delete-qzdmf\" (UID: \"6873cb35-815f-44ae-b38d-bb7ceb806b9d\") " pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.806744 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.812710 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.863311 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.887601 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-jw42k"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.890980 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkz9s\" (UniqueName: \"kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s\") pod \"novaapic192-account-delete-qzdmf\" (UID: \"6873cb35-815f-44ae-b38d-bb7ceb806b9d\") " pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.891996 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-jw42k"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.904312 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt4xp\" (UniqueName: \"kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp\") pod \"novacell11e55-account-delete-zrnhb\" (UID: \"5a3bb280-8572-448d-a3af-4c33993fc232\") " pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.908238 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.921875 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.922169 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="dnsmasq-dns" containerID="cri-o://b5ac45277657c7bc51337c63b0014ecaf12e5118db0d75abef1c824653bddcaa" gracePeriod=10 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.977560 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.983608 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7cf5d6fffb-62nql" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-log" containerID="cri-o://0526bde28972f0419714df8a5e921f83607729d39715b762ffd0a2c5dd4a7841" gracePeriod=30 Oct 01 07:29:12 crc kubenswrapper[4837]: I1001 07:29:12.984130 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7cf5d6fffb-62nql" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-api" containerID="cri-o://5f3d01a41efebb2a65c4e6e279d47ec4c2349d13a91fa0e7c788a05bf6c23dbd" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.014291 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt4xp\" (UniqueName: \"kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp\") pod \"novacell11e55-account-delete-zrnhb\" (UID: \"5a3bb280-8572-448d-a3af-4c33993fc232\") " pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.014781 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: E1001 07:29:13.014866 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:13 crc kubenswrapper[4837]: E1001 07:29:13.014922 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data podName:96e5feff-74ce-434a-8606-85c51844deb9 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:15.01490537 +0000 UTC m=+1411.856512825 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data") pod "rabbitmq-cell1-server-0" (UID: "96e5feff-74ce-434a-8606-85c51844deb9") : configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.015063 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="cinder-scheduler" containerID="cri-o://73ae4948895720b7496ecb1982bf5fbab9c712bead101852e3e795d3237de601" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.015482 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="probe" containerID="cri-o://bf54d4f5f766e15a07eac31a84eb955d37b5b5dfce2d9dbf7292f0ae70e3db3f" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.034208 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-tvk89"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.078633 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-tvk89"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.098894 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt4xp\" (UniqueName: \"kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp\") pod \"novacell11e55-account-delete-zrnhb\" (UID: \"5a3bb280-8572-448d-a3af-4c33993fc232\") " pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.144353 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee312865-a3af-43d5-9b7d-13d62695af31/ovsdbserver-nb/0.log" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.144752 4837 generic.go:334] "Generic (PLEG): container finished" podID="ee312865-a3af-43d5-9b7d-13d62695af31" containerID="a33be5ffcfa7d1c7be08f25ec05a88d909c111fd9dde1c9328ba84acd170fc4c" exitCode=2 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.144775 4837 generic.go:334] "Generic (PLEG): container finished" podID="ee312865-a3af-43d5-9b7d-13d62695af31" containerID="b2dd01f2720f0c29fc64551b59dc951b915dcc75366896c51b0ec3db97f15550" exitCode=143 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.144885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerDied","Data":"a33be5ffcfa7d1c7be08f25ec05a88d909c111fd9dde1c9328ba84acd170fc4c"} Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.144921 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerDied","Data":"b2dd01f2720f0c29fc64551b59dc951b915dcc75366896c51b0ec3db97f15550"} Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.172034 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.172966 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api-log" containerID="cri-o://c5de51b5f6c2a76c7e0d181f7cdfb7392fdc8323c0f38c8d9eaf84c7dfe54255" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.173383 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api" containerID="cri-o://e8db386829c99e752736e8a3a56fdc0ddce5b550d7fd78fdcfb1d074bd8ca628" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.182018 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.182216 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cf464dc69-n54w7" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-api" containerID="cri-o://59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.182520 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cf464dc69-n54w7" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-httpd" containerID="cri-o://2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.193819 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_144c538f-4172-4043-b716-62f91b64f8fc/ovsdbserver-sb/0.log" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.193892 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194270 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194638 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-server" containerID="cri-o://6d2afe280ee355f22575011affdf80ede81294951985ed4fdfaf0a0a8bd04577" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194721 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="swift-recon-cron" containerID="cri-o://4eb16d26f4116a6a74e9d0b4accc3991ac0da36440e5760b73b34784878eb561" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194752 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="rsync" containerID="cri-o://0bfa759f888d0212de7ee4b0712ad5d33add96370a9084085bce043b2d375d44" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194778 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-expirer" containerID="cri-o://2067fe40fd9f9d75412077801cde319e8068868f481690bf6f201f36f8f71642" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194803 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-updater" containerID="cri-o://708ad62022e8e219ec2f8ff1ce8f2f1e128a8e14443aec6458580876f7636171" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194829 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-auditor" containerID="cri-o://fd642928ddcba69b2aa30d3fdad7ae984f220ec2171638957d4a6f7678a20b45" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194860 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-replicator" containerID="cri-o://06276e79f10f7789a399e46f0f9ba7ceccd3ca7926331f03e09147648c3966d4" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194887 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-server" containerID="cri-o://80729d7b80fe7ca0a76deb60aded53df6ae5bafa464acedacb113c4fcfa46c02" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194936 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-updater" containerID="cri-o://b5626cefa6e75a0a68d30120dccd151dc40d2c9a0f69f966c98d8c027ef04148" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194973 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-auditor" containerID="cri-o://dbb2c5bf1f5499683d8d50b2b442d96eef02c58748d98bc5851c45b9a40621df" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.194999 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-replicator" containerID="cri-o://feb371a16a28f5d9590fbbaf472765ac149570124d961c2b6244deaac26355e5" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195025 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-server" containerID="cri-o://9290d642faa7786d6b897ad12a0754d4d61afe7b651852d4dab287495dddc287" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195049 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-reaper" containerID="cri-o://2204c01a3c6692934f14277b4d5cbf6cca7a5c501820bb3fa3389ae18fea6f76" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195075 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-auditor" containerID="cri-o://8caed28b92cf49fa4a285edcda33f9590fa05d856cd61e8bcce897cc1725a303" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195103 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-replicator" containerID="cri-o://c67bf25539192ca5553d8c61e702cf3adae4424f5c481f14196a5218a1c14dde" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195447 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2vw77_28035a28-28d4-46b3-b269-97a523cf731e/openstack-network-exporter/0.log" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195476 4837 generic.go:334] "Generic (PLEG): container finished" podID="28035a28-28d4-46b3-b269-97a523cf731e" containerID="fa33ac993198b8bd4e6712d1dfc94ce0a3b46c83683d716c97d0c41883173646" exitCode=2 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.195518 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2vw77" event={"ID":"28035a28-28d4-46b3-b269-97a523cf731e","Type":"ContainerDied","Data":"fa33ac993198b8bd4e6712d1dfc94ce0a3b46c83683d716c97d0c41883173646"} Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.217921 4837 generic.go:334] "Generic (PLEG): container finished" podID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerID="a699f84318a8fa1ea08dd47672464162bcd065e0a5b23d594adac792bb355f6b" exitCode=2 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.217968 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerDied","Data":"a699f84318a8fa1ea08dd47672464162bcd065e0a5b23d594adac792bb355f6b"} Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.226084 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:13 crc kubenswrapper[4837]: E1001 07:29:13.271310 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 01 07:29:13 crc kubenswrapper[4837]: E1001 07:29:13.271403 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data podName:0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:15.271385739 +0000 UTC m=+1412.112993184 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data") pod "rabbitmq-server-0" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087") : configmap "rabbitmq-config-data" not found Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.362435 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bphz9"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379401 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379459 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379574 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379672 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48msk\" (UniqueName: \"kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379702 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379754 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379789 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.379823 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs\") pod \"144c538f-4172-4043-b716-62f91b64f8fc\" (UID: \"144c538f-4172-4043-b716-62f91b64f8fc\") " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.406932 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bphz9"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.409778 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts" (OuterVolumeSpecName: "scripts") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.412328 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.420963 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk" (OuterVolumeSpecName: "kube-api-access-48msk") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "kube-api-access-48msk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.433576 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config" (OuterVolumeSpecName: "config") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.446065 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: connect: connection refused" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.446365 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.581475 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.581504 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/144c538f-4172-4043-b716-62f91b64f8fc-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.581535 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.581571 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48msk\" (UniqueName: \"kubernetes.io/projected/144c538f-4172-4043-b716-62f91b64f8fc-kube-api-access-48msk\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.581583 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144c538f-4172-4043-b716-62f91b64f8fc-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.603612 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-l5h6t"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.674266 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-l5h6t"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.686861 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.714126 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-kjn5l"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.747377 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-kjn5l"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.758411 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.761253 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e1a9-account-create-6drcx"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.769583 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e1a9-account-create-6drcx"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.780314 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.780375 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.791627 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.791653 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.814290 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="rabbitmq" containerID="cri-o://3f3ea1fd9b8bc426d4733c4c8e9a603415784a810c094ebe41057d96f5b75c2e" gracePeriod=604800 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.814381 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.846022 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d5a60a-28ff-4ce9-858e-e6835a77b65f" path="/var/lib/kubelet/pods/08d5a60a-28ff-4ce9-858e-e6835a77b65f/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.851997 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="201dd153-cadb-48ca-bddc-b7ba0622f39a" path="/var/lib/kubelet/pods/201dd153-cadb-48ca-bddc-b7ba0622f39a/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.852854 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303dfe99-b3ef-4f14-857b-014fd13dce75" path="/var/lib/kubelet/pods/303dfe99-b3ef-4f14-857b-014fd13dce75/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.854947 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88" path="/var/lib/kubelet/pods/7f4a0fc8-f0d7-4bc4-9c3a-d53421e88e88/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.856668 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbf0ac2-1b84-4e30-8328-31554d960caf" path="/var/lib/kubelet/pods/7fbf0ac2-1b84-4e30-8328-31554d960caf/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.859027 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80f48dbc-6a43-46a7-bca1-00027ded7d1f" path="/var/lib/kubelet/pods/80f48dbc-6a43-46a7-bca1-00027ded7d1f/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.869291 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ca37a0-8611-42e6-b365-7ed2aa8dc811" path="/var/lib/kubelet/pods/a1ca37a0-8611-42e6-b365-7ed2aa8dc811/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.870207 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacbfd69-af34-4267-8e75-eee0aba34fe9" path="/var/lib/kubelet/pods/cacbfd69-af34-4267-8e75-eee0aba34fe9/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.870697 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1c94750-373c-41d4-aa27-7d580f35b967" path="/var/lib/kubelet/pods/d1c94750-373c-41d4-aa27-7d580f35b967/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.876643 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0af960-55d3-436a-b532-f8460076932e" path="/var/lib/kubelet/pods/fc0af960-55d3-436a-b532-f8460076932e/volumes" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.883354 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.883448 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.886669 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-log" containerID="cri-o://4db0657ecb1d6bab9cfa0aca6d6c4da781d4dc1788c9788ff50b8ac94c614f0d" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.887130 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-httpd" containerID="cri-o://a708b9492c517db3c4ac65ac6d8e06a0899c2e2382102fdd99be860d6cabafbe" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.900443 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.905677 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "144c538f-4172-4043-b716-62f91b64f8fc" (UID: "144c538f-4172-4043-b716-62f91b64f8fc"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.955441 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2vw77_28035a28-28d4-46b3-b269-97a523cf731e/openstack-network-exporter/0.log" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.955522 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.975469 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.975770 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-log" containerID="cri-o://b1a25a5631787393515c8e82608962228bfb05a6623f191895b95414994acc1b" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.976196 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-httpd" containerID="cri-o://470c5f7992ab6223b3d21e3422fa5783032efa7c901479e728e382c4209680ac" gracePeriod=30 Oct 01 07:29:13 crc kubenswrapper[4837]: I1001 07:29:13.990784 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.008100 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": read tcp 10.217.0.2:38214->10.217.0.173:9292: read: connection reset by peer" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.010419 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.010615 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.012042 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config" (OuterVolumeSpecName: "config") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.018494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.018586 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.018635 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.018699 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgzkk\" (UniqueName: \"kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk\") pod \"28035a28-28d4-46b3-b269-97a523cf731e\" (UID: \"28035a28-28d4-46b3-b269-97a523cf731e\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.018805 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.019139 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.020287 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.020350 4837 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28035a28-28d4-46b3-b269-97a523cf731e-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.020432 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28035a28-28d4-46b3-b269-97a523cf731e-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.020483 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/144c538f-4172-4043-b716-62f91b64f8fc-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.027957 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk" (OuterVolumeSpecName: "kube-api-access-zgzkk") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "kube-api-access-zgzkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.044827 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-rh7lh"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.045983 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee312865-a3af-43d5-9b7d-13d62695af31/ovsdbserver-nb/0.log" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.046426 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.065731 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-rh7lh"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.107028 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="rabbitmq" containerID="cri-o://afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f" gracePeriod=604800 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.119609 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122430 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122562 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122599 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqtz4\" (UniqueName: \"kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122655 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122753 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122780 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122801 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.122870 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs\") pod \"ee312865-a3af-43d5-9b7d-13d62695af31\" (UID: \"ee312865-a3af-43d5-9b7d-13d62695af31\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.123243 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgzkk\" (UniqueName: \"kubernetes.io/projected/28035a28-28d4-46b3-b269-97a523cf731e-kube-api-access-zgzkk\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.129181 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" containerID="cri-o://64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" gracePeriod=29 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.130673 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config" (OuterVolumeSpecName: "config") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.135316 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.135400 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.137658 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts" (OuterVolumeSpecName: "scripts") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.145737 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4" (OuterVolumeSpecName: "kube-api-access-fqtz4") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "kube-api-access-fqtz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: E1001 07:29:14.146220 4837 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 01 07:29:14 crc kubenswrapper[4837]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 01 07:29:14 crc kubenswrapper[4837]: + source /usr/local/bin/container-scripts/functions Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNBridge=br-int Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNRemote=tcp:localhost:6642 Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNEncapType=geneve Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNAvailabilityZones= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ EnableChassisAsGateway=true Oct 01 07:29:14 crc kubenswrapper[4837]: ++ PhysicalNetworks= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNHostName= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 01 07:29:14 crc kubenswrapper[4837]: ++ ovs_dir=/var/lib/openvswitch Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 01 07:29:14 crc kubenswrapper[4837]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + sleep 0.5 Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + cleanup_ovsdb_server_semaphore Oct 01 07:29:14 crc kubenswrapper[4837]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 01 07:29:14 crc kubenswrapper[4837]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-zpx7h" message=< Oct 01 07:29:14 crc kubenswrapper[4837]: Exiting ovsdb-server (5) [ OK ] Oct 01 07:29:14 crc kubenswrapper[4837]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 01 07:29:14 crc kubenswrapper[4837]: + source /usr/local/bin/container-scripts/functions Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNBridge=br-int Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNRemote=tcp:localhost:6642 Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNEncapType=geneve Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNAvailabilityZones= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ EnableChassisAsGateway=true Oct 01 07:29:14 crc kubenswrapper[4837]: ++ PhysicalNetworks= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNHostName= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 01 07:29:14 crc kubenswrapper[4837]: ++ ovs_dir=/var/lib/openvswitch Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 01 07:29:14 crc kubenswrapper[4837]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + sleep 0.5 Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + cleanup_ovsdb_server_semaphore Oct 01 07:29:14 crc kubenswrapper[4837]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 01 07:29:14 crc kubenswrapper[4837]: > Oct 01 07:29:14 crc kubenswrapper[4837]: E1001 07:29:14.146286 4837 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 01 07:29:14 crc kubenswrapper[4837]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 01 07:29:14 crc kubenswrapper[4837]: + source /usr/local/bin/container-scripts/functions Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNBridge=br-int Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNRemote=tcp:localhost:6642 Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNEncapType=geneve Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNAvailabilityZones= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ EnableChassisAsGateway=true Oct 01 07:29:14 crc kubenswrapper[4837]: ++ PhysicalNetworks= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ OVNHostName= Oct 01 07:29:14 crc kubenswrapper[4837]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 01 07:29:14 crc kubenswrapper[4837]: ++ ovs_dir=/var/lib/openvswitch Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 01 07:29:14 crc kubenswrapper[4837]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 01 07:29:14 crc kubenswrapper[4837]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + sleep 0.5 Oct 01 07:29:14 crc kubenswrapper[4837]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 01 07:29:14 crc kubenswrapper[4837]: + cleanup_ovsdb_server_semaphore Oct 01 07:29:14 crc kubenswrapper[4837]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 01 07:29:14 crc kubenswrapper[4837]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 01 07:29:14 crc kubenswrapper[4837]: > pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" containerID="cri-o://d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.146325 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" containerID="cri-o://d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" gracePeriod=29 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.174038 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8fcf-account-create-t728v"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.177713 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.190639 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-mdbmd"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.198410 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-mdbmd"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.205176 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8fcf-account-create-t728v"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.215799 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.225593 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.225800 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqtz4\" (UniqueName: \"kubernetes.io/projected/ee312865-a3af-43d5-9b7d-13d62695af31-kube-api-access-fqtz4\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.225878 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.225967 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.226040 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee312865-a3af-43d5-9b7d-13d62695af31-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.226111 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.231296 4837 generic.go:334] "Generic (PLEG): container finished" podID="73e65e7c-ced2-4e38-9841-3f383471b275" containerID="b1a25a5631787393515c8e82608962228bfb05a6623f191895b95414994acc1b" exitCode=143 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.231364 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerDied","Data":"b1a25a5631787393515c8e82608962228bfb05a6623f191895b95414994acc1b"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.231861 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-sj7hz"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.241848 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2vw77_28035a28-28d4-46b3-b269-97a523cf731e/openstack-network-exporter/0.log" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.241919 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2vw77" event={"ID":"28035a28-28d4-46b3-b269-97a523cf731e","Type":"ContainerDied","Data":"7d44d925e70e92de59c9cd1cf5d9b55b40f013297d40d8cd8e72b2d5b659418d"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.241954 4837 scope.go:117] "RemoveContainer" containerID="fa33ac993198b8bd4e6712d1dfc94ce0a3b46c83683d716c97d0c41883173646" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.242118 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2vw77" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.252629 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6315-account-create-rb6zg"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258095 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="0bfa759f888d0212de7ee4b0712ad5d33add96370a9084085bce043b2d375d44" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258123 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="2067fe40fd9f9d75412077801cde319e8068868f481690bf6f201f36f8f71642" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258133 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="708ad62022e8e219ec2f8ff1ce8f2f1e128a8e14443aec6458580876f7636171" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258146 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="fd642928ddcba69b2aa30d3fdad7ae984f220ec2171638957d4a6f7678a20b45" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258156 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="06276e79f10f7789a399e46f0f9ba7ceccd3ca7926331f03e09147648c3966d4" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258164 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="80729d7b80fe7ca0a76deb60aded53df6ae5bafa464acedacb113c4fcfa46c02" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258172 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="b5626cefa6e75a0a68d30120dccd151dc40d2c9a0f69f966c98d8c027ef04148" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258183 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="dbb2c5bf1f5499683d8d50b2b442d96eef02c58748d98bc5851c45b9a40621df" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258192 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="feb371a16a28f5d9590fbbaf472765ac149570124d961c2b6244deaac26355e5" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258200 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="9290d642faa7786d6b897ad12a0754d4d61afe7b651852d4dab287495dddc287" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258209 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="2204c01a3c6692934f14277b4d5cbf6cca7a5c501820bb3fa3389ae18fea6f76" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258217 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="8caed28b92cf49fa4a285edcda33f9590fa05d856cd61e8bcce897cc1725a303" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258225 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="c67bf25539192ca5553d8c61e702cf3adae4424f5c481f14196a5218a1c14dde" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258233 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="6d2afe280ee355f22575011affdf80ede81294951985ed4fdfaf0a0a8bd04577" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"0bfa759f888d0212de7ee4b0712ad5d33add96370a9084085bce043b2d375d44"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258302 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"2067fe40fd9f9d75412077801cde319e8068868f481690bf6f201f36f8f71642"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258316 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"708ad62022e8e219ec2f8ff1ce8f2f1e128a8e14443aec6458580876f7636171"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258329 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"fd642928ddcba69b2aa30d3fdad7ae984f220ec2171638957d4a6f7678a20b45"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258342 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"06276e79f10f7789a399e46f0f9ba7ceccd3ca7926331f03e09147648c3966d4"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258354 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"80729d7b80fe7ca0a76deb60aded53df6ae5bafa464acedacb113c4fcfa46c02"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258365 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"b5626cefa6e75a0a68d30120dccd151dc40d2c9a0f69f966c98d8c027ef04148"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258376 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"dbb2c5bf1f5499683d8d50b2b442d96eef02c58748d98bc5851c45b9a40621df"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258392 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"feb371a16a28f5d9590fbbaf472765ac149570124d961c2b6244deaac26355e5"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"9290d642faa7786d6b897ad12a0754d4d61afe7b651852d4dab287495dddc287"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258420 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"2204c01a3c6692934f14277b4d5cbf6cca7a5c501820bb3fa3389ae18fea6f76"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258431 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"8caed28b92cf49fa4a285edcda33f9590fa05d856cd61e8bcce897cc1725a303"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258442 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"c67bf25539192ca5553d8c61e702cf3adae4424f5c481f14196a5218a1c14dde"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.258453 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"6d2afe280ee355f22575011affdf80ede81294951985ed4fdfaf0a0a8bd04577"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.262371 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_144c538f-4172-4043-b716-62f91b64f8fc/ovsdbserver-sb/0.log" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.262435 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"144c538f-4172-4043-b716-62f91b64f8fc","Type":"ContainerDied","Data":"0a85e84c711ab07e9e92ee43b138119648100285a03a48d1d7e39d9624c83e1a"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.262541 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.285761 4837 scope.go:117] "RemoveContainer" containerID="167c29409de05ee8281b1aed4f4b8f38acfffec5ceb180aebcf6ac61a6a016c5" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.292712 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-sj7hz"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.296540 4837 generic.go:334] "Generic (PLEG): container finished" podID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerID="c5de51b5f6c2a76c7e0d181f7cdfb7392fdc8323c0f38c8d9eaf84c7dfe54255" exitCode=143 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.297368 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerDied","Data":"c5de51b5f6c2a76c7e0d181f7cdfb7392fdc8323c0f38c8d9eaf84c7dfe54255"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.301276 4837 generic.go:334] "Generic (PLEG): container finished" podID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerID="b5ac45277657c7bc51337c63b0014ecaf12e5118db0d75abef1c824653bddcaa" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.301425 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.301431 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" event={"ID":"34a848f4-7b85-4463-bdae-f763e6f9e7e5","Type":"ContainerDied","Data":"b5ac45277657c7bc51337c63b0014ecaf12e5118db0d75abef1c824653bddcaa"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.305025 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindere1a9-account-delete-qw9xn" event={"ID":"8f72fdf5-3656-49e2-80a8-37d0d65ec212","Type":"ContainerStarted","Data":"2545e1d9b3763a3d5c9ad129e708b52d1b9ff318d6745a3c04a20f8b10ed51c5"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.306457 4837 generic.go:334] "Generic (PLEG): container finished" podID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" containerID="c5da1cb2cb85875baa579d1f9fd8e420c616437d114859950730be4875ac535e" exitCode=137 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.306640 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.312982 4837 generic.go:334] "Generic (PLEG): container finished" podID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerID="4db0657ecb1d6bab9cfa0aca6d6c4da781d4dc1788c9788ff50b8ac94c614f0d" exitCode=143 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.313128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerDied","Data":"4db0657ecb1d6bab9cfa0aca6d6c4da781d4dc1788c9788ff50b8ac94c614f0d"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.313881 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.315682 4837 generic.go:334] "Generic (PLEG): container finished" podID="6212cb23-af83-4bcf-8328-91b14a62844b" containerID="0526bde28972f0419714df8a5e921f83607729d39715b762ffd0a2c5dd4a7841" exitCode=143 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.315813 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerDied","Data":"0526bde28972f0419714df8a5e921f83607729d39715b762ffd0a2c5dd4a7841"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.318565 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.318988 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-869997d995-swqv9" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-httpd" containerID="cri-o://c8bdd89e1b73c7441649e0ed8d304c53f3b5ed251d4c5fd8292eb1d03c32ff83" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.319535 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-869997d995-swqv9" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-server" containerID="cri-o://5d1ccc4bac17d9e48187ea2dd17c1341921ed55167a7aeb5bfba6f2afe624a09" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.322439 4837 generic.go:334] "Generic (PLEG): container finished" podID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerID="2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.322573 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerDied","Data":"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.326126 4837 generic.go:334] "Generic (PLEG): container finished" podID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" exitCode=0 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.326294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerDied","Data":"d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.327972 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.333991 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6315-account-create-rb6zg"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.335802 4837 scope.go:117] "RemoveContainer" containerID="48c9eecb63371cb6c6437405cb59dd74a49403ac0f542fdf6ce9200b6b395b32" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.342706 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f4b1-account-create-lzfzs"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.349614 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee312865-a3af-43d5-9b7d-13d62695af31/ovsdbserver-nb/0.log" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.349744 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee312865-a3af-43d5-9b7d-13d62695af31","Type":"ContainerDied","Data":"ed403b18b72ad5bfcd988a39cf2a3cbcc8a9c28a2048534a8fa96bb821df130b"} Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.349896 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.356914 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.365451 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f4b1-account-create-lzfzs"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.380619 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.384653 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.388122 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.388408 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-log" containerID="cri-o://2422f422ed0b438a27d2a54ccb39f04357a6213eac4f80a06db39e4b2ebad2be" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.388549 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-api" containerID="cri-o://e4a74db66f683da3b67436f59d8f19271ded325f5539b6a7b61f313bce201a1c" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.393606 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-869997d995-swqv9" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.156:8080/healthcheck\": dial tcp 10.217.0.156:8080: connect: connection refused" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.395424 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.397186 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-869997d995-swqv9" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.156:8080/healthcheck\": dial tcp 10.217.0.156:8080: connect: connection refused" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.420686 4837 scope.go:117] "RemoveContainer" containerID="c5da1cb2cb85875baa579d1f9fd8e420c616437d114859950730be4875ac535e" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.423187 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.427067 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" containerID="cri-o://d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.428226 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" containerID="cri-o://121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.429722 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config\") pod \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.429770 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.429883 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwwxj\" (UniqueName: \"kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430014 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430056 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430132 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w67zv\" (UniqueName: \"kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv\") pod \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430184 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430249 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle\") pod \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430317 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret\") pod \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\" (UID: \"4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.430342 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb\") pod \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\" (UID: \"34a848f4-7b85-4463-bdae-f763e6f9e7e5\") " Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.431653 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.492059 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj" (OuterVolumeSpecName: "kube-api-access-xwwxj") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "kube-api-access-xwwxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.495959 4837 scope.go:117] "RemoveContainer" containerID="a33be5ffcfa7d1c7be08f25ec05a88d909c111fd9dde1c9328ba84acd170fc4c" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.501605 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv" (OuterVolumeSpecName: "kube-api-access-w67zv") pod "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" (UID: "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc"). InnerVolumeSpecName "kube-api-access-w67zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.507102 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-fbd8-account-create-5wlnj"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.527373 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-fbd8-account-create-5wlnj"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.537299 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwwxj\" (UniqueName: \"kubernetes.io/projected/34a848f4-7b85-4463-bdae-f763e6f9e7e5-kube-api-access-xwwxj\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.537328 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w67zv\" (UniqueName: \"kubernetes.io/projected/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-kube-api-access-w67zv\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.539761 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1e55-account-create-7588j"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.545376 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1e55-account-create-7588j"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.552598 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.553053 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66dc5bc7b8-ttchd" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api-log" containerID="cri-o://b7792f2f321c19d1eb1568ae96eddc87b77661b3d811ceac5cfe3f464d5b9726" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.553463 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66dc5bc7b8-ttchd" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api" containerID="cri-o://9d0c63a56b505c323247613cb05d39bfe8227df3c32c1ff6ebcabfa1e90233e6" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.564624 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-skz66"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.582077 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.582567 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener-log" containerID="cri-o://44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.582935 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener" containerID="cri-o://e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: W1001 07:29:14.603064 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode273fd30_371c_494e_99ce_6460b2a1dd00.slice/crio-758e9fc071b4e96ed478a1adbcc56981884cdf43ae091ea0e77b53eadd9334fa WatchSource:0}: Error finding container 758e9fc071b4e96ed478a1adbcc56981884cdf43ae091ea0e77b53eadd9334fa: Status 404 returned error can't find the container with id 758e9fc071b4e96ed478a1adbcc56981884cdf43ae091ea0e77b53eadd9334fa Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.603160 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.612060 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-skz66"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.621930 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jd57t"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.633297 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jd57t"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.644119 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.644359 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68b46954cf-msjzx" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker-log" containerID="cri-o://4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.644799 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68b46954cf-msjzx" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker" containerID="cri-o://3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.669792 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "28035a28-28d4-46b3-b269-97a523cf731e" (UID: "28035a28-28d4-46b3-b269-97a523cf731e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.682594 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.683503 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ddfefe16-d728-4570-9ef6-cc717162f590" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.706355 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.706546 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerName="nova-cell1-conductor-conductor" containerID="cri-o://ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.729096 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.756531 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.756565 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28035a28-28d4-46b3-b269-97a523cf731e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.759229 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ee312865-a3af-43d5-9b7d-13d62695af31" (UID: "ee312865-a3af-43d5-9b7d-13d62695af31"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.763243 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lrdgf"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.768563 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.778413 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lrdgf"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.788647 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vd4tb"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.813542 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.814074 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" containerName="nova-cell0-conductor-conductor" containerID="cri-o://5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" gracePeriod=30 Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.851743 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" (UID: "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.855225 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vd4tb"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.861758 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee312865-a3af-43d5-9b7d-13d62695af31-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.878497 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.878597 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.864003 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" (UID: "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.868242 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.925146 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.935212 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.956192 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config" (OuterVolumeSpecName: "config") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.968227 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34a848f4-7b85-4463-bdae-f763e6f9e7e5" (UID: "34a848f4-7b85-4463-bdae-f763e6f9e7e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.973672 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.981010 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.981041 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.981051 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.981059 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:14 crc kubenswrapper[4837]: I1001 07:29:14.981068 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a848f4-7b85-4463-bdae-f763e6f9e7e5-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.012627 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.024959 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" (UID: "4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.030518 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.046916 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.058810 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.059028 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerName="nova-scheduler-scheduler" containerID="cri-o://2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" gracePeriod=30 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.066774 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.076820 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.080573 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.082483 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.082560 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.082609 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data podName:96e5feff-74ce-434a-8606-85c51844deb9 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:19.082591948 +0000 UTC m=+1415.924199403 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data") pod "rabbitmq-cell1-server-0" (UID: "96e5feff-74ce-434a-8606-85c51844deb9") : configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.082798 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="galera" containerID="cri-o://c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223" gracePeriod=30 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.085659 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.169167 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.182178 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-2vw77"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.213940 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.238139 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.289883 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.289953 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data podName:0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:19.289938478 +0000 UTC m=+1416.131545933 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data") pod "rabbitmq-server-0" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087") : configmap "rabbitmq-config-data" not found Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.308888 4837 scope.go:117] "RemoveContainer" containerID="b2dd01f2720f0c29fc64551b59dc951b915dcc75366896c51b0ec3db97f15550" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.379445 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell11e55-account-delete-zrnhb" event={"ID":"5a3bb280-8572-448d-a3af-4c33993fc232","Type":"ContainerStarted","Data":"5a64151aab7eba9497b930d27759203672dcc22c39ac279ea15ba0e70dcff67a"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.388035 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic192-account-delete-qzdmf" event={"ID":"6873cb35-815f-44ae-b38d-bb7ceb806b9d","Type":"ContainerStarted","Data":"0914ae8ca47f0a9342bbeb91f2ddb67b624fec7e9654b08e871f232252327823"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.401447 4837 generic.go:334] "Generic (PLEG): container finished" podID="8f72fdf5-3656-49e2-80a8-37d0d65ec212" containerID="d82ff67647eae0dc78cadc5398b75190de9934067eddeef92311560e0c706fd1" exitCode=0 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.401570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindere1a9-account-delete-qw9xn" event={"ID":"8f72fdf5-3656-49e2-80a8-37d0d65ec212","Type":"ContainerDied","Data":"d82ff67647eae0dc78cadc5398b75190de9934067eddeef92311560e0c706fd1"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.406545 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanf4b1-account-delete-x448d" event={"ID":"aa5e54c4-ad98-4f26-9681-189997f3bd64","Type":"ContainerStarted","Data":"55bf00143d371180808b2001da232a23239eddc8a9027b50463b6aeaa36fc53e"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.410197 4837 generic.go:334] "Generic (PLEG): container finished" podID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerID="4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144" exitCode=143 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.410244 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerDied","Data":"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.430933 4837 generic.go:334] "Generic (PLEG): container finished" podID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerID="b7792f2f321c19d1eb1568ae96eddc87b77661b3d811ceac5cfe3f464d5b9726" exitCode=143 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.431096 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerDied","Data":"b7792f2f321c19d1eb1568ae96eddc87b77661b3d811ceac5cfe3f464d5b9726"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.444097 4837 generic.go:334] "Generic (PLEG): container finished" podID="6063ad95-cbaf-4798-899f-195d4428f246" containerID="d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a" exitCode=143 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.444204 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerDied","Data":"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.453050 4837 generic.go:334] "Generic (PLEG): container finished" podID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerID="2422f422ed0b438a27d2a54ccb39f04357a6213eac4f80a06db39e4b2ebad2be" exitCode=143 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.453162 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerDied","Data":"2422f422ed0b438a27d2a54ccb39f04357a6213eac4f80a06db39e4b2ebad2be"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.464041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement7d99-account-delete-pnxgc" event={"ID":"dce4369c-e009-4ae1-bc07-65dff0005807","Type":"ContainerStarted","Data":"9df226eeaa4878d1f21fffa90c8b0f057e8a1d28ae4ba39973ea9ef3833672fc"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.471926 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6315-account-delete-ddhmt" event={"ID":"595b1839-8525-4b91-ab30-fd800710cf31","Type":"ContainerStarted","Data":"d8c190e33adde7954694055619eab50e0af752af97028a89bcf0b0312c3da7a1"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.482917 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" event={"ID":"34a848f4-7b85-4463-bdae-f763e6f9e7e5","Type":"ContainerDied","Data":"37138bfce86319542c491e929d2e04e41ee31e25ba2a9d815395bd4458116854"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.483001 4837 scope.go:117] "RemoveContainer" containerID="b5ac45277657c7bc51337c63b0014ecaf12e5118db0d75abef1c824653bddcaa" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.483011 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.491664 4837 generic.go:334] "Generic (PLEG): container finished" podID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerID="bf54d4f5f766e15a07eac31a84eb955d37b5b5dfce2d9dbf7292f0ae70e3db3f" exitCode=0 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.491706 4837 generic.go:334] "Generic (PLEG): container finished" podID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerID="73ae4948895720b7496ecb1982bf5fbab9c712bead101852e3e795d3237de601" exitCode=0 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.491746 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerDied","Data":"bf54d4f5f766e15a07eac31a84eb955d37b5b5dfce2d9dbf7292f0ae70e3db3f"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.491768 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerDied","Data":"73ae4948895720b7496ecb1982bf5fbab9c712bead101852e3e795d3237de601"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.523429 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron8fcf-account-delete-g5pcd" event={"ID":"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8","Type":"ContainerStarted","Data":"c5baf46f726032871dd64ccd1f5b4476799f4d0a86660295880c383122e5ce99"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.526533 4837 generic.go:334] "Generic (PLEG): container finished" podID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerID="44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666" exitCode=143 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.526582 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerDied","Data":"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.528269 4837 generic.go:334] "Generic (PLEG): container finished" podID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerID="5d1ccc4bac17d9e48187ea2dd17c1341921ed55167a7aeb5bfba6f2afe624a09" exitCode=0 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.528295 4837 generic.go:334] "Generic (PLEG): container finished" podID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerID="c8bdd89e1b73c7441649e0ed8d304c53f3b5ed251d4c5fd8292eb1d03c32ff83" exitCode=0 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.528326 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerDied","Data":"5d1ccc4bac17d9e48187ea2dd17c1341921ed55167a7aeb5bfba6f2afe624a09"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.528343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerDied","Data":"c8bdd89e1b73c7441649e0ed8d304c53f3b5ed251d4c5fd8292eb1d03c32ff83"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.529648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerStarted","Data":"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.529672 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerStarted","Data":"758e9fc071b4e96ed478a1adbcc56981884cdf43ae091ea0e77b53eadd9334fa"} Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.550995 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.632284 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.634035 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.634171 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.635310 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.635387 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.636437 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.638773 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.638794 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.836947 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="144c538f-4172-4043-b716-62f91b64f8fc" path="/var/lib/kubelet/pods/144c538f-4172-4043-b716-62f91b64f8fc/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.837551 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28035a28-28d4-46b3-b269-97a523cf731e" path="/var/lib/kubelet/pods/28035a28-28d4-46b3-b269-97a523cf731e/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.838182 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39965847-66b7-45ad-b24e-1ce81e39515b" path="/var/lib/kubelet/pods/39965847-66b7-45ad-b24e-1ce81e39515b/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.839579 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4420e1fc-5656-453c-b5a4-31d5d7fe0012" path="/var/lib/kubelet/pods/4420e1fc-5656-453c-b5a4-31d5d7fe0012/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.840105 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc" path="/var/lib/kubelet/pods/4ca6b254-cf4b-435a-a9b1-d6fa63d64dbc/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.840548 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c35a796-0dbb-41de-b8c3-222c4dcd1ef5" path="/var/lib/kubelet/pods/7c35a796-0dbb-41de-b8c3-222c4dcd1ef5/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.841043 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9681804a-3e4e-44a1-b1fc-8280c44806bd" path="/var/lib/kubelet/pods/9681804a-3e4e-44a1-b1fc-8280c44806bd/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.846154 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abc5c46e-5e15-4c96-b1d8-7867068c488d" path="/var/lib/kubelet/pods/abc5c46e-5e15-4c96-b1d8-7867068c488d/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.847313 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef375f9-54f9-4c27-88c8-8a8345d4a5f8" path="/var/lib/kubelet/pods/aef375f9-54f9-4c27-88c8-8a8345d4a5f8/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.849313 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c4895a-6bd7-4bd6-9beb-bb61bd04b970" path="/var/lib/kubelet/pods/b3c4895a-6bd7-4bd6-9beb-bb61bd04b970/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.852012 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf665e4-7264-4c3b-875e-84ef7aeedd96" path="/var/lib/kubelet/pods/cdf665e4-7264-4c3b-875e-84ef7aeedd96/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.853241 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20e80bb-46a9-4b04-b2d3-132683af9fb3" path="/var/lib/kubelet/pods/d20e80bb-46a9-4b04-b2d3-132683af9fb3/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.855568 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8322b3d-5eac-45bc-ad35-f98e62cf9715" path="/var/lib/kubelet/pods/d8322b3d-5eac-45bc-ad35-f98e62cf9715/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.857108 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb" path="/var/lib/kubelet/pods/dcc652ae-7ae1-473f-9fa9-eb2e26d9efdb/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.860549 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba7e445-3312-44a0-ab5e-7cdd14676b93" path="/var/lib/kubelet/pods/eba7e445-3312-44a0-ab5e-7cdd14676b93/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.862256 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" path="/var/lib/kubelet/pods/ee312865-a3af-43d5-9b7d-13d62695af31/volumes" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.871874 4837 scope.go:117] "RemoveContainer" containerID="9d47745a8b132f2982a19fd66ee8dfb226b79b906c64bbaffadbf5953e32dcb4" Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.914377 4837 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 01 07:29:15 crc kubenswrapper[4837]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-01T07:29:13Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 01 07:29:15 crc kubenswrapper[4837]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Oct 01 07:29:15 crc kubenswrapper[4837]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-kt6jr" message=< Oct 01 07:29:15 crc kubenswrapper[4837]: Exiting ovn-controller (1) [FAILED] Oct 01 07:29:15 crc kubenswrapper[4837]: Killing ovn-controller (1) [ OK ] Oct 01 07:29:15 crc kubenswrapper[4837]: 2025-10-01T07:29:13Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 01 07:29:15 crc kubenswrapper[4837]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Oct 01 07:29:15 crc kubenswrapper[4837]: > Oct 01 07:29:15 crc kubenswrapper[4837]: E1001 07:29:15.914412 4837 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 01 07:29:15 crc kubenswrapper[4837]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-01T07:29:13Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 01 07:29:15 crc kubenswrapper[4837]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Oct 01 07:29:15 crc kubenswrapper[4837]: > pod="openstack/ovn-controller-kt6jr" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" containerID="cri-o://b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.914459 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-kt6jr" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" containerID="cri-o://b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" gracePeriod=27 Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.921219 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kt6jr" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" probeResult="failure" output="" Oct 01 07:29:15 crc kubenswrapper[4837]: I1001 07:29:15.972743 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114424 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fm9r\" (UniqueName: \"kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114515 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114583 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114615 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114683 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.114777 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom\") pod \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\" (UID: \"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.116061 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.121216 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts" (OuterVolumeSpecName: "scripts") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.121636 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.123359 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r" (OuterVolumeSpecName: "kube-api-access-9fm9r") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "kube-api-access-9fm9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.151129 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.160934 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.171797 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.217219 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fm9r\" (UniqueName: \"kubernetes.io/projected/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-kube-api-access-9fm9r\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.217244 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.217253 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.217263 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.294316 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kt6jr_295f86b5-4cd2-4ceb-bf30-b3397a6f5f62/ovn-controller/0.log" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.294405 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318389 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtwcf\" (UniqueName: \"kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf\") pod \"ddfefe16-d728-4570-9ef6-cc717162f590\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318442 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318484 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlnwg\" (UniqueName: \"kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg\") pod \"8f72fdf5-3656-49e2-80a8-37d0d65ec212\" (UID: \"8f72fdf5-3656-49e2-80a8-37d0d65ec212\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318511 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318530 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle\") pod \"ddfefe16-d728-4570-9ef6-cc717162f590\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318554 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318576 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318607 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data\") pod \"ddfefe16-d728-4570-9ef6-cc717162f590\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318650 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs\") pod \"ddfefe16-d728-4570-9ef6-cc717162f590\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318683 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8q4z\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318752 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318815 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs\") pod \"ddfefe16-d728-4570-9ef6-cc717162f590\" (UID: \"ddfefe16-d728-4570-9ef6-cc717162f590\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318852 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.318873 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs\") pod \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\" (UID: \"9d39fc78-2eb7-4a57-8bbc-11733cbb366a\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.328080 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.332880 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.333676 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.335291 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-njhjj"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.341394 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-njhjj"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.355211 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.362058 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.368718 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7d99-account-create-grnjj"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.374089 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7d99-account-create-grnjj"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.376800 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z" (OuterVolumeSpecName: "kube-api-access-h8q4z") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "kube-api-access-h8q4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.377257 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg" (OuterVolumeSpecName: "kube-api-access-wlnwg") pod "8f72fdf5-3656-49e2-80a8-37d0d65ec212" (UID: "8f72fdf5-3656-49e2-80a8-37d0d65ec212"). InnerVolumeSpecName "kube-api-access-wlnwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.379593 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.384512 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf" (OuterVolumeSpecName: "kube-api-access-dtwcf") pod "ddfefe16-d728-4570-9ef6-cc717162f590" (UID: "ddfefe16-d728-4570-9ef6-cc717162f590"). InnerVolumeSpecName "kube-api-access-dtwcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420508 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420558 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97dld\" (UniqueName: \"kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420593 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420631 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420720 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.421837 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422479 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.420854 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422553 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422618 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422715 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle\") pod \"44374bf9-c936-4c06-ac74-e80244c09039\" (UID: \"44374bf9-c936-4c06-ac74-e80244c09039\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422732 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422766 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tbcv\" (UniqueName: \"kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.422791 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts\") pod \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\" (UID: \"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62\") " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423651 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtwcf\" (UniqueName: \"kubernetes.io/projected/ddfefe16-d728-4570-9ef6-cc717162f590-kube-api-access-dtwcf\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423662 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423671 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlnwg\" (UniqueName: \"kubernetes.io/projected/8f72fdf5-3656-49e2-80a8-37d0d65ec212-kube-api-access-wlnwg\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423679 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423690 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423709 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423718 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8q4z\" (UniqueName: \"kubernetes.io/projected/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-kube-api-access-h8q4z\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.423726 4837 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.424654 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts" (OuterVolumeSpecName: "scripts") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.424689 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.425471 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run" (OuterVolumeSpecName: "var-run") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.426217 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.426478 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.427981 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.442234 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets" (OuterVolumeSpecName: "secrets") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.443429 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld" (OuterVolumeSpecName: "kube-api-access-97dld") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "kube-api-access-97dld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.443913 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv" (OuterVolumeSpecName: "kube-api-access-8tbcv") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "kube-api-access-8tbcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.476846 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526137 4837 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526239 4837 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44374bf9-c936-4c06-ac74-e80244c09039-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526348 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526438 4837 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526512 4837 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526566 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tbcv\" (UniqueName: \"kubernetes.io/projected/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-kube-api-access-8tbcv\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526636 4837 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526726 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526809 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97dld\" (UniqueName: \"kubernetes.io/projected/44374bf9-c936-4c06-ac74-e80244c09039-kube-api-access-97dld\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.526882 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/44374bf9-c936-4c06-ac74-e80244c09039-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.549460 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-869997d995-swqv9" event={"ID":"9d39fc78-2eb7-4a57-8bbc-11733cbb366a","Type":"ContainerDied","Data":"d76ca20d8e415d77e2049d146b84bc4e6beda0554fdc6547178b387ec6891ccd"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.549512 4837 scope.go:117] "RemoveContainer" containerID="5d1ccc4bac17d9e48187ea2dd17c1341921ed55167a7aeb5bfba6f2afe624a09" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.549608 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-869997d995-swqv9" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.558265 4837 generic.go:334] "Generic (PLEG): container finished" podID="dce4369c-e009-4ae1-bc07-65dff0005807" containerID="f2e57a498d2d7002b6196b4a7d582862dc279abc76d6c4f29681768a4d3c6a6f" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.558351 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement7d99-account-delete-pnxgc" event={"ID":"dce4369c-e009-4ae1-bc07-65dff0005807","Type":"ContainerDied","Data":"f2e57a498d2d7002b6196b4a7d582862dc279abc76d6c4f29681768a4d3c6a6f"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.566773 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.573388 4837 generic.go:334] "Generic (PLEG): container finished" podID="6873cb35-815f-44ae-b38d-bb7ceb806b9d" containerID="a5007423f20f31799bd3b83421d45c3a51a7e25c84dc7deaaee06262d96de1ef" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.573475 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic192-account-delete-qzdmf" event={"ID":"6873cb35-815f-44ae-b38d-bb7ceb806b9d","Type":"ContainerDied","Data":"a5007423f20f31799bd3b83421d45c3a51a7e25c84dc7deaaee06262d96de1ef"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.583478 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70ff50da-fcbe-4fe9-9a35-073f2abdc3e7","Type":"ContainerDied","Data":"d9c89ed9bfc11964e0d5b30ae091d4dc0f4c1763318f924641f2df54697e6807"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.583561 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.588081 4837 generic.go:334] "Generic (PLEG): container finished" podID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerID="e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.588138 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerDied","Data":"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.616883 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.617857 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.628125 4837 generic.go:334] "Generic (PLEG): container finished" podID="6212cb23-af83-4bcf-8328-91b14a62844b" containerID="5f3d01a41efebb2a65c4e6e279d47ec4c2349d13a91fa0e7c788a05bf6c23dbd" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.628182 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerDied","Data":"5f3d01a41efebb2a65c4e6e279d47ec4c2349d13a91fa0e7c788a05bf6c23dbd"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.628299 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.628328 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.628343 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.633568 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "ddfefe16-d728-4570-9ef6-cc717162f590" (UID: "ddfefe16-d728-4570-9ef6-cc717162f590"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.640219 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell11e55-account-delete-zrnhb" event={"ID":"5a3bb280-8572-448d-a3af-4c33993fc232","Type":"ContainerStarted","Data":"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.643932 4837 generic.go:334] "Generic (PLEG): container finished" podID="ddfefe16-d728-4570-9ef6-cc717162f590" containerID="4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.643971 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddfefe16-d728-4570-9ef6-cc717162f590","Type":"ContainerDied","Data":"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.643989 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddfefe16-d728-4570-9ef6-cc717162f590","Type":"ContainerDied","Data":"2edc6f515b97e4a41b981ab02aa6490b9169ab572df9b27cfe7a7c6ac4513ba2"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.644029 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.645790 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kt6jr_295f86b5-4cd2-4ceb-bf30-b3397a6f5f62/ovn-controller/0.log" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.645812 4837 generic.go:334] "Generic (PLEG): container finished" podID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerID="b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" exitCode=143 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.645845 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr" event={"ID":"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62","Type":"ContainerDied","Data":"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.645860 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kt6jr" event={"ID":"295f86b5-4cd2-4ceb-bf30-b3397a6f5f62","Type":"ContainerDied","Data":"c847295e489a5acc261e47c3c1388e5e6983c1f8d86bc2a53ea898868e3d233d"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.645897 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kt6jr" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.648234 4837 generic.go:334] "Generic (PLEG): container finished" podID="44374bf9-c936-4c06-ac74-e80244c09039" containerID="c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.648265 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerDied","Data":"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.648279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"44374bf9-c936-4c06-ac74-e80244c09039","Type":"ContainerDied","Data":"fc4268e9129deeb80bfeefd6bd5daedba90027177c86e100dc825a945316fd84"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.648393 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.654578 4837 generic.go:334] "Generic (PLEG): container finished" podID="aa5e54c4-ad98-4f26-9681-189997f3bd64" containerID="0c61265ca20b5940cf1844ccd442b04f5de4c5d17b8de5b7bfbba2dd3bbb9d3c" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.654649 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanf4b1-account-delete-x448d" event={"ID":"aa5e54c4-ad98-4f26-9681-189997f3bd64","Type":"ContainerDied","Data":"0c61265ca20b5940cf1844ccd442b04f5de4c5d17b8de5b7bfbba2dd3bbb9d3c"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.660754 4837 generic.go:334] "Generic (PLEG): container finished" podID="2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" containerID="0e91c05a507a757b672ef0ff4aa99493b289fd25a5e7eb5cf9c6c892390a95bf" exitCode=0 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.660822 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron8fcf-account-delete-g5pcd" event={"ID":"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8","Type":"ContainerDied","Data":"0e91c05a507a757b672ef0ff4aa99493b289fd25a5e7eb5cf9c6c892390a95bf"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.669528 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindere1a9-account-delete-qw9xn" event={"ID":"8f72fdf5-3656-49e2-80a8-37d0d65ec212","Type":"ContainerDied","Data":"2545e1d9b3763a3d5c9ad129e708b52d1b9ff318d6745a3c04a20f8b10ed51c5"} Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.670320 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindere1a9-account-delete-qw9xn" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.678868 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.699351 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.711153 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "ddfefe16-d728-4570-9ef6-cc717162f590" (UID: "ddfefe16-d728-4570-9ef6-cc717162f590"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.718328 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddfefe16-d728-4570-9ef6-cc717162f590" (UID: "ddfefe16-d728-4570-9ef6-cc717162f590"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.730653 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.730683 4837 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.730715 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.730725 4837 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.761351 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.761658 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data" (OuterVolumeSpecName: "config-data") pod "ddfefe16-d728-4570-9ef6-cc717162f590" (UID: "ddfefe16-d728-4570-9ef6-cc717162f590"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.804034 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data" (OuterVolumeSpecName: "config-data") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.820460 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "44374bf9-c936-4c06-ac74-e80244c09039" (UID: "44374bf9-c936-4c06-ac74-e80244c09039"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.840655 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.840687 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfefe16-d728-4570-9ef6-cc717162f590-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.840710 4837 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/44374bf9-c936-4c06-ac74-e80244c09039-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.840720 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.849118 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.849360 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-central-agent" containerID="cri-o://ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.850026 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="proxy-httpd" containerID="cri-o://6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.850077 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="sg-core" containerID="cri-o://50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.850107 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-notification-agent" containerID="cri-o://2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.867204 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" (UID: "295f86b5-4cd2-4ceb-bf30-b3397a6f5f62"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.882656 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.888842 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b2851153-f304-4041-9de2-045c0d92526d" containerName="kube-state-metrics" containerID="cri-o://034af4cc06f98a5a1c1bf69b7c633186fde48f42cccb9b972d748a605cd9eec8" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.937795 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.941977 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.942005 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.953294 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data" (OuterVolumeSpecName: "config-data") pod "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" (UID: "70ff50da-fcbe-4fe9-9a35-073f2abdc3e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.983794 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.983994 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" containerName="memcached" containerID="cri-o://afc5fac4c4bdbcdfcb537ec4258dde192ccbd7e0b4e6a075117de4ead1c2405d" gracePeriod=30 Oct 01 07:29:16 crc kubenswrapper[4837]: I1001 07:29:16.986098 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9d39fc78-2eb7-4a57-8bbc-11733cbb366a" (UID: "9d39fc78-2eb7-4a57-8bbc-11733cbb366a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:16 crc kubenswrapper[4837]: E1001 07:29:16.991346 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.014397 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.049983 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39fc78-2eb7-4a57-8bbc-11733cbb366a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.050013 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.067002 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.067239 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerName="nova-cell1-conductor-conductor" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.079259 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-pn9tb"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.101921 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-pn9tb"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.117540 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fpv6z"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.135673 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fpv6z"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.157104 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.157353 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5799b746f7-mzft5" podUID="943a976f-044a-4b42-b08c-baeae7b85f42" containerName="keystone-api" containerID="cri-o://6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3" gracePeriod=30 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.162414 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.173:9292/healthcheck\": read tcp 10.217.0.2:38210->10.217.0.173:9292: read: connection reset by peer" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.191755 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.265955 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-fnmv2"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.322112 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-fnmv2"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.346301 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-53a7-account-create-gc7dc"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.357456 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.358602 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-53a7-account-create-gc7dc"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.386861 4837 scope.go:117] "RemoveContainer" containerID="c8bdd89e1b73c7441649e0ed8d304c53f3b5ed251d4c5fd8292eb1d03c32ff83" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.399814 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2w99m"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.406410 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.406657 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.413268 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2w99m"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.421403 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.429831 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.429903 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.452937 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.467392 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.467495 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.470868 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.470911 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.471020 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.471053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqxs9\" (UniqueName: \"kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.471121 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs\") pod \"6212cb23-af83-4bcf-8328-91b14a62844b\" (UID: \"6212cb23-af83-4bcf-8328-91b14a62844b\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.471170 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c192-account-create-52hlt"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.471346 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="galera" containerID="cri-o://2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" gracePeriod=30 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.472051 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs" (OuterVolumeSpecName: "logs") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.479760 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.483052 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts" (OuterVolumeSpecName: "scripts") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.498191 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9" (OuterVolumeSpecName: "kube-api-access-lqxs9") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "kube-api-access-lqxs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.501689 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.507787 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c192-account-create-52hlt"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.512275 4837 scope.go:117] "RemoveContainer" containerID="bf54d4f5f766e15a07eac31a84eb955d37b5b5dfce2d9dbf7292f0ae70e3db3f" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.526198 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.547410 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.556038 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cindere1a9-account-delete-qw9xn"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.560077 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.568854 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.572729 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qkl7\" (UniqueName: \"kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7\") pod \"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8\" (UID: \"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.572925 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkkt7\" (UniqueName: \"kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7\") pod \"aa5e54c4-ad98-4f26-9681-189997f3bd64\" (UID: \"aa5e54c4-ad98-4f26-9681-189997f3bd64\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.572976 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-689ks\" (UniqueName: \"kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks\") pod \"dce4369c-e009-4ae1-bc07-65dff0005807\" (UID: \"dce4369c-e009-4ae1-bc07-65dff0005807\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.573426 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqxs9\" (UniqueName: \"kubernetes.io/projected/6212cb23-af83-4bcf-8328-91b14a62844b-kube-api-access-lqxs9\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.573445 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6212cb23-af83-4bcf-8328-91b14a62844b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.573453 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.573875 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data" (OuterVolumeSpecName: "config-data") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.589405 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.596759 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kt6jr"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.597439 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7" (OuterVolumeSpecName: "kube-api-access-vkkt7") pod "aa5e54c4-ad98-4f26-9681-189997f3bd64" (UID: "aa5e54c4-ad98-4f26-9681-189997f3bd64"). InnerVolumeSpecName "kube-api-access-vkkt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.600143 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7" (OuterVolumeSpecName: "kube-api-access-6qkl7") pod "2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" (UID: "2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8"). InnerVolumeSpecName "kube-api-access-6qkl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.605685 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.607894 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks" (OuterVolumeSpecName: "kube-api-access-689ks") pod "dce4369c-e009-4ae1-bc07-65dff0005807" (UID: "dce4369c-e009-4ae1-bc07-65dff0005807"). InnerVolumeSpecName "kube-api-access-689ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.607943 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.616085 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.617504 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.647545 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:55722->10.217.0.200:8775: read: connection reset by peer" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.647572 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:55734->10.217.0.200:8775: read: connection reset by peer" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.649971 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.659831 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-869997d995-swqv9"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.675869 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt4xp\" (UniqueName: \"kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp\") pod \"5a3bb280-8572-448d-a3af-4c33993fc232\" (UID: \"5a3bb280-8572-448d-a3af-4c33993fc232\") " Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.676379 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-689ks\" (UniqueName: \"kubernetes.io/projected/dce4369c-e009-4ae1-bc07-65dff0005807-kube-api-access-689ks\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.676392 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qkl7\" (UniqueName: \"kubernetes.io/projected/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8-kube-api-access-6qkl7\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.676401 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.676411 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkkt7\" (UniqueName: \"kubernetes.io/projected/aa5e54c4-ad98-4f26-9681-189997f3bd64-kube-api-access-vkkt7\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.696866 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp" (OuterVolumeSpecName: "kube-api-access-qt4xp") pod "5a3bb280-8572-448d-a3af-4c33993fc232" (UID: "5a3bb280-8572-448d-a3af-4c33993fc232"). InnerVolumeSpecName "kube-api-access-qt4xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.708812 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement7d99-account-delete-pnxgc" event={"ID":"dce4369c-e009-4ae1-bc07-65dff0005807","Type":"ContainerDied","Data":"9df226eeaa4878d1f21fffa90c8b0f057e8a1d28ae4ba39973ea9ef3833672fc"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.708907 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement7d99-account-delete-pnxgc" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.741388 4837 generic.go:334] "Generic (PLEG): container finished" podID="595b1839-8525-4b91-ab30-fd800710cf31" containerID="942fd5c680f88e7c4fd1e2e59bf8058b4e347d103f7e8b089ff68968743b0e69" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.741443 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6315-account-delete-ddhmt" event={"ID":"595b1839-8525-4b91-ab30-fd800710cf31","Type":"ContainerDied","Data":"942fd5c680f88e7c4fd1e2e59bf8058b4e347d103f7e8b089ff68968743b0e69"} Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.744642 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.751447 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanf4b1-account-delete-x448d" event={"ID":"aa5e54c4-ad98-4f26-9681-189997f3bd64","Type":"ContainerDied","Data":"55bf00143d371180808b2001da232a23239eddc8a9027b50463b6aeaa36fc53e"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.751554 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanf4b1-account-delete-x448d" Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.758891 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.767225 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088 is running failed: container process not found" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 07:29:17 crc kubenswrapper[4837]: E1001 07:29:17.767279 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerName="nova-scheduler-scheduler" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.769143 4837 generic.go:334] "Generic (PLEG): container finished" podID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerID="ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.769188 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b","Type":"ContainerDied","Data":"ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.777554 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt4xp\" (UniqueName: \"kubernetes.io/projected/5a3bb280-8572-448d-a3af-4c33993fc232-kube-api-access-qt4xp\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.777864 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerStarted","Data":"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.780689 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7cf5d6fffb-62nql" event={"ID":"6212cb23-af83-4bcf-8328-91b14a62844b","Type":"ContainerDied","Data":"acea766de68aecf54adb5656fbbed042c79ac7bbd52ffa173d7469fa3356ae48"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.780800 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7cf5d6fffb-62nql" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.782482 4837 generic.go:334] "Generic (PLEG): container finished" podID="5a3bb280-8572-448d-a3af-4c33993fc232" containerID="ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b" exitCode=1 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.782522 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell11e55-account-delete-zrnhb" event={"ID":"5a3bb280-8572-448d-a3af-4c33993fc232","Type":"ContainerDied","Data":"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.782536 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell11e55-account-delete-zrnhb" event={"ID":"5a3bb280-8572-448d-a3af-4c33993fc232","Type":"ContainerDied","Data":"5a64151aab7eba9497b930d27759203672dcc22c39ac279ea15ba0e70dcff67a"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.782584 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell11e55-account-delete-zrnhb" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.837743 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.848435 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.848614 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0416426e-d7dc-458e-aa3a-e1c4a160757d" path="/var/lib/kubelet/pods/0416426e-d7dc-458e-aa3a-e1c4a160757d/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.849334 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fa0495-6203-43f0-aacb-534dcb9a4b04" path="/var/lib/kubelet/pods/11fa0495-6203-43f0-aacb-534dcb9a4b04/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.849874 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" path="/var/lib/kubelet/pods/295f86b5-4cd2-4ceb-bf30-b3397a6f5f62/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.852115 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": read tcp 10.217.0.2:34712->10.217.0.165:8776: read: connection reset by peer" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.857391 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44374bf9-c936-4c06-ac74-e80244c09039" path="/var/lib/kubelet/pods/44374bf9-c936-4c06-ac74-e80244c09039/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.861396 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a67cce-6fe6-4bf8-9f10-36eb0544e8b7" path="/var/lib/kubelet/pods/50a67cce-6fe6-4bf8-9f10-36eb0544e8b7/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.861893 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" path="/var/lib/kubelet/pods/70ff50da-fcbe-4fe9-9a35-073f2abdc3e7/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.862967 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85661570-740d-45f3-a34c-bc575512ca64" path="/var/lib/kubelet/pods/85661570-740d-45f3-a34c-bc575512ca64/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.863397 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f72fdf5-3656-49e2-80a8-37d0d65ec212" path="/var/lib/kubelet/pods/8f72fdf5-3656-49e2-80a8-37d0d65ec212/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.863932 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99acf906-a3d0-4aeb-9fc8-3c9f9ba28654" path="/var/lib/kubelet/pods/99acf906-a3d0-4aeb-9fc8-3c9f9ba28654/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.865801 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" path="/var/lib/kubelet/pods/9d39fc78-2eb7-4a57-8bbc-11733cbb366a/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.866657 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c" path="/var/lib/kubelet/pods/9d8e10dd-a7a8-49ce-9fa3-a4a63d02366c/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.876554 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23ccb73-ce30-474a-9939-1adc6bde5d67" path="/var/lib/kubelet/pods/c23ccb73-ce30-474a-9939-1adc6bde5d67/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.877335 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddfefe16-d728-4570-9ef6-cc717162f590" path="/var/lib/kubelet/pods/ddfefe16-d728-4570-9ef6-cc717162f590/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.878277 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6212cb23-af83-4bcf-8328-91b14a62844b" (UID: "6212cb23-af83-4bcf-8328-91b14a62844b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.879109 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.879125 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.879134 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6212cb23-af83-4bcf-8328-91b14a62844b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.879527 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f64b4342-3448-432d-88e7-7c182524aa19" path="/var/lib/kubelet/pods/f64b4342-3448-432d-88e7-7c182524aa19/volumes" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.903584 4837 generic.go:334] "Generic (PLEG): container finished" podID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerID="e4a74db66f683da3b67436f59d8f19271ded325f5539b6a7b61f313bce201a1c" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.904958 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerDied","Data":"e4a74db66f683da3b67436f59d8f19271ded325f5539b6a7b61f313bce201a1c"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.926076 4837 generic.go:334] "Generic (PLEG): container finished" podID="b2851153-f304-4041-9de2-045c0d92526d" containerID="034af4cc06f98a5a1c1bf69b7c633186fde48f42cccb9b972d748a605cd9eec8" exitCode=2 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.926163 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2851153-f304-4041-9de2-045c0d92526d","Type":"ContainerDied","Data":"034af4cc06f98a5a1c1bf69b7c633186fde48f42cccb9b972d748a605cd9eec8"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.926193 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2851153-f304-4041-9de2-045c0d92526d","Type":"ContainerDied","Data":"47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.926203 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47d8d7a9a753cccbeb70e378197bde081be02c6ae05e27991206b6542fa53e0f" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.934884 4837 generic.go:334] "Generic (PLEG): container finished" podID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerID="6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.934912 4837 generic.go:334] "Generic (PLEG): container finished" podID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerID="50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a" exitCode=2 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.934920 4837 generic.go:334] "Generic (PLEG): container finished" podID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerID="ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.934978 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerDied","Data":"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.935000 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerDied","Data":"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.935012 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerDied","Data":"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.943242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron8fcf-account-delete-g5pcd" event={"ID":"2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8","Type":"ContainerDied","Data":"c5baf46f726032871dd64ccd1f5b4476799f4d0a86660295880c383122e5ce99"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.943337 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron8fcf-account-delete-g5pcd" Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.956566 4837 generic.go:334] "Generic (PLEG): container finished" podID="73e65e7c-ced2-4e38-9841-3f383471b275" containerID="470c5f7992ab6223b3d21e3422fa5783032efa7c901479e728e382c4209680ac" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.956624 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerDied","Data":"470c5f7992ab6223b3d21e3422fa5783032efa7c901479e728e382c4209680ac"} Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.967521 4837 generic.go:334] "Generic (PLEG): container finished" podID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerID="a708b9492c517db3c4ac65ac6d8e06a0899c2e2382102fdd99be860d6cabafbe" exitCode=0 Oct 01 07:29:17 crc kubenswrapper[4837]: I1001 07:29:17.967755 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerDied","Data":"a708b9492c517db3c4ac65ac6d8e06a0899c2e2382102fdd99be860d6cabafbe"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.256484 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.261680 4837 scope.go:117] "RemoveContainer" containerID="73ae4948895720b7496ecb1982bf5fbab9c712bead101852e3e795d3237de601" Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.293479 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.296975 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.298642 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.299402 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.299461 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="galera" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.309310 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.333860 4837 scope.go:117] "RemoveContainer" containerID="4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.366964 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.378324 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.396102 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config\") pod \"b2851153-f304-4041-9de2-045c0d92526d\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.396179 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.396198 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.396216 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.397083 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.396325 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398004 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m68mq\" (UniqueName: \"kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398029 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6flkz\" (UniqueName: \"kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz\") pod \"b2851153-f304-4041-9de2-045c0d92526d\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398051 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398065 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle\") pod \"b2851153-f304-4041-9de2-045c0d92526d\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398092 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398130 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398157 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398200 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm4sj\" (UniqueName: \"kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398237 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398253 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398271 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398312 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs\") pod \"23126d10-0ad7-45e9-94e3-3e34116c20d8\" (UID: \"23126d10-0ad7-45e9-94e3-3e34116c20d8\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398340 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398364 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle\") pod \"73e65e7c-ced2-4e38-9841-3f383471b275\" (UID: \"73e65e7c-ced2-4e38-9841-3f383471b275\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398389 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs\") pod \"b2851153-f304-4041-9de2-045c0d92526d\" (UID: \"b2851153-f304-4041-9de2-045c0d92526d\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.398804 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.401555 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.408098 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.411234 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs" (OuterVolumeSpecName: "logs") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.411833 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs" (OuterVolumeSpecName: "logs") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.412298 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.421158 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq" (OuterVolumeSpecName: "kube-api-access-m68mq") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "kube-api-access-m68mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.424887 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.430152 4837 scope.go:117] "RemoveContainer" containerID="4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.430271 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj" (OuterVolumeSpecName: "kube-api-access-xm4sj") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "kube-api-access-xm4sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.462507 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.462831 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e\": container with ID starting with 4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e not found: ID does not exist" containerID="4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.462874 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e"} err="failed to get container status \"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e\": rpc error: code = NotFound desc = could not find container \"4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e\": container with ID starting with 4e583d8a3095c2e563a6c5a7702fdb17b86d881c0a4d5877a99084cccbd2785e not found: ID does not exist" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.462900 4837 scope.go:117] "RemoveContainer" containerID="b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.465290 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.480190 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz" (OuterVolumeSpecName: "kube-api-access-6flkz") pod "b2851153-f304-4041-9de2-045c0d92526d" (UID: "b2851153-f304-4041-9de2-045c0d92526d"). InnerVolumeSpecName "kube-api-access-6flkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.482002 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts" (OuterVolumeSpecName: "scripts") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.482083 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts" (OuterVolumeSpecName: "scripts") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.488283 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement7d99-account-delete-pnxgc"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.497474 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle\") pod \"6063ad95-cbaf-4798-899f-195d4428f246\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499568 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t446f\" (UniqueName: \"kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f\") pod \"595b1839-8525-4b91-ab30-fd800710cf31\" (UID: \"595b1839-8525-4b91-ab30-fd800710cf31\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499617 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data\") pod \"6063ad95-cbaf-4798-899f-195d4428f246\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499657 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data\") pod \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499709 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data\") pod \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499746 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t962t\" (UniqueName: \"kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t\") pod \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle\") pod \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\" (UID: \"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499786 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs\") pod \"6063ad95-cbaf-4798-899f-195d4428f246\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499859 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle\") pod \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499889 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5bx9\" (UniqueName: \"kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9\") pod \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\" (UID: \"7f15a72f-f4ec-4e07-ae6d-7eef5700f543\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499911 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs\") pod \"6063ad95-cbaf-4798-899f-195d4428f246\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.499959 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzrnb\" (UniqueName: \"kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb\") pod \"6063ad95-cbaf-4798-899f-195d4428f246\" (UID: \"6063ad95-cbaf-4798-899f-195d4428f246\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500247 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m68mq\" (UniqueName: \"kubernetes.io/projected/23126d10-0ad7-45e9-94e3-3e34116c20d8-kube-api-access-m68mq\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500259 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6flkz\" (UniqueName: \"kubernetes.io/projected/b2851153-f304-4041-9de2-045c0d92526d-kube-api-access-6flkz\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500267 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500275 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500283 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm4sj\" (UniqueName: \"kubernetes.io/projected/73e65e7c-ced2-4e38-9841-3f383471b275-kube-api-access-xm4sj\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500291 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500299 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23126d10-0ad7-45e9-94e3-3e34116c20d8-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500307 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73e65e7c-ced2-4e38-9841-3f383471b275-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500323 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.500335 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.508597 4837 scope.go:117] "RemoveContainer" containerID="b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.510842 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs" (OuterVolumeSpecName: "logs") pod "6063ad95-cbaf-4798-899f-195d4428f246" (UID: "6063ad95-cbaf-4798-899f-195d4428f246"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: E1001 07:29:18.510875 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef\": container with ID starting with b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef not found: ID does not exist" containerID="b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.511128 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef"} err="failed to get container status \"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef\": rpc error: code = NotFound desc = could not find container \"b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef\": container with ID starting with b1c284ccc577e3b63e78c362769a37667b734aad6cfe66bf642c11f37557b8ef not found: ID does not exist" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.511264 4837 scope.go:117] "RemoveContainer" containerID="c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.529023 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7cf5d6fffb-62nql"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.530239 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.540188 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.541475 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f" (OuterVolumeSpecName: "kube-api-access-t446f") pod "595b1839-8525-4b91-ab30-fd800710cf31" (UID: "595b1839-8525-4b91-ab30-fd800710cf31"). InnerVolumeSpecName "kube-api-access-t446f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.543946 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanf4b1-account-delete-x448d"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.545313 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.553785 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.558594 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "b2851153-f304-4041-9de2-045c0d92526d" (UID: "b2851153-f304-4041-9de2-045c0d92526d"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.558865 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb" (OuterVolumeSpecName: "kube-api-access-zzrnb") pod "6063ad95-cbaf-4798-899f-195d4428f246" (UID: "6063ad95-cbaf-4798-899f-195d4428f246"). InnerVolumeSpecName "kube-api-access-zzrnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.560208 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron8fcf-account-delete-g5pcd"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.568755 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.569608 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t" (OuterVolumeSpecName: "kube-api-access-t962t") pod "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" (UID: "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b"). InnerVolumeSpecName "kube-api-access-t962t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.573342 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9" (OuterVolumeSpecName: "kube-api-access-k5bx9") pod "7f15a72f-f4ec-4e07-ae6d-7eef5700f543" (UID: "7f15a72f-f4ec-4e07-ae6d-7eef5700f543"). InnerVolumeSpecName "kube-api-access-k5bx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.574923 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell11e55-account-delete-zrnhb"] Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601125 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601259 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601286 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601307 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601342 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601382 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601421 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601439 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqzlh\" (UniqueName: \"kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601522 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601571 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601614 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601635 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdj6t\" (UniqueName: \"kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601676 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd\") pod \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\" (UID: \"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.601815 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle\") pod \"08552fbf-b050-46ec-b8bf-83cde04f87e6\" (UID: \"08552fbf-b050-46ec-b8bf-83cde04f87e6\") " Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602146 4837 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602166 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t962t\" (UniqueName: \"kubernetes.io/projected/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-kube-api-access-t962t\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602176 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5bx9\" (UniqueName: \"kubernetes.io/projected/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-kube-api-access-k5bx9\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602184 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6063ad95-cbaf-4798-899f-195d4428f246-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602194 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzrnb\" (UniqueName: \"kubernetes.io/projected/6063ad95-cbaf-4798-899f-195d4428f246-kube-api-access-zzrnb\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602203 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t446f\" (UniqueName: \"kubernetes.io/projected/595b1839-8525-4b91-ab30-fd800710cf31-kube-api-access-t446f\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.602989 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.603853 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs" (OuterVolumeSpecName: "logs") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.616816 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh" (OuterVolumeSpecName: "kube-api-access-jqzlh") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "kube-api-access-jqzlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.617611 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.626858 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts" (OuterVolumeSpecName: "scripts") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.629379 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t" (OuterVolumeSpecName: "kube-api-access-vdj6t") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "kube-api-access-vdj6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.674957 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703668 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703722 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqzlh\" (UniqueName: \"kubernetes.io/projected/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-kube-api-access-jqzlh\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703735 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08552fbf-b050-46ec-b8bf-83cde04f87e6-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703744 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdj6t\" (UniqueName: \"kubernetes.io/projected/08552fbf-b050-46ec-b8bf-83cde04f87e6-kube-api-access-vdj6t\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703753 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703761 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.703771 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.707280 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6063ad95-cbaf-4798-899f-195d4428f246" (UID: "6063ad95-cbaf-4798-899f-195d4428f246"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.711155 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "b2851153-f304-4041-9de2-045c0d92526d" (UID: "b2851153-f304-4041-9de2-045c0d92526d"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.722848 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data" (OuterVolumeSpecName: "config-data") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.740941 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.749841 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.752416 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.773671 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f15a72f-f4ec-4e07-ae6d-7eef5700f543" (UID: "7f15a72f-f4ec-4e07-ae6d-7eef5700f543"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.778907 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6063ad95-cbaf-4798-899f-195d4428f246" (UID: "6063ad95-cbaf-4798-899f-195d4428f246"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.789297 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" (UID: "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.795660 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.800652 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816574 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816623 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816634 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816644 4837 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816655 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816665 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816672 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816711 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816720 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816728 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.816737 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.824256 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data" (OuterVolumeSpecName: "config-data") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.826785 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data" (OuterVolumeSpecName: "config-data") pod "6063ad95-cbaf-4798-899f-195d4428f246" (UID: "6063ad95-cbaf-4798-899f-195d4428f246"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.828277 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data" (OuterVolumeSpecName: "config-data") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.828289 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data" (OuterVolumeSpecName: "config-data") pod "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" (UID: "c46a0ffe-6c41-4d34-bbe9-8db505bbe46b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.828278 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.828339 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.828885 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2851153-f304-4041-9de2-045c0d92526d" (UID: "b2851153-f304-4041-9de2-045c0d92526d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.835329 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data" (OuterVolumeSpecName: "config-data") pod "7f15a72f-f4ec-4e07-ae6d-7eef5700f543" (UID: "7f15a72f-f4ec-4e07-ae6d-7eef5700f543"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.836083 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "73e65e7c-ced2-4e38-9841-3f383471b275" (UID: "73e65e7c-ced2-4e38-9841-3f383471b275"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.854517 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.864718 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "08552fbf-b050-46ec-b8bf-83cde04f87e6" (UID: "08552fbf-b050-46ec-b8bf-83cde04f87e6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.867749 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "23126d10-0ad7-45e9-94e3-3e34116c20d8" (UID: "23126d10-0ad7-45e9-94e3-3e34116c20d8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.887880 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data" (OuterVolumeSpecName: "config-data") pod "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" (UID: "17575b0a-a27e-4cc5-a84a-0ed9adf0ed65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918371 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2851153-f304-4041-9de2-045c0d92526d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918397 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918407 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918415 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918425 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918433 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918441 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73e65e7c-ced2-4e38-9841-3f383471b275-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918451 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/08552fbf-b050-46ec-b8bf-83cde04f87e6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918458 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23126d10-0ad7-45e9-94e3-3e34116c20d8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918466 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6063ad95-cbaf-4798-899f-195d4428f246-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918475 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918482 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.918490 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f15a72f-f4ec-4e07-ae6d-7eef5700f543-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.985540 4837 generic.go:334] "Generic (PLEG): container finished" podID="6063ad95-cbaf-4798-899f-195d4428f246" containerID="121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191" exitCode=0 Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.985582 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.985612 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerDied","Data":"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.985655 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6063ad95-cbaf-4798-899f-195d4428f246","Type":"ContainerDied","Data":"2c968f6fbc476f7bf90bae775aa43cc0ec58edfa101b4488fdb46bf73a31717b"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.989049 4837 generic.go:334] "Generic (PLEG): container finished" podID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" exitCode=0 Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.989092 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f15a72f-f4ec-4e07-ae6d-7eef5700f543","Type":"ContainerDied","Data":"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.989118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f15a72f-f4ec-4e07-ae6d-7eef5700f543","Type":"ContainerDied","Data":"82ebaffe8662e75b736bd799fabae478db52300184689fa90da793dc63211dd4"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.989165 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.994239 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"73e65e7c-ced2-4e38-9841-3f383471b275","Type":"ContainerDied","Data":"57b012ebf793a59f0c6135f174c2d9d51b48b236c83964bd864f5cfd18979864"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.994321 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.999394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"08552fbf-b050-46ec-b8bf-83cde04f87e6","Type":"ContainerDied","Data":"c420ce38f93c3996d8ce993dda04baa9ba1172bb04b4072544170d6e860a8507"} Oct 01 07:29:18 crc kubenswrapper[4837]: I1001 07:29:18.999465 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.009514 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c46a0ffe-6c41-4d34-bbe9-8db505bbe46b","Type":"ContainerDied","Data":"276a680b19af91fda764eff2425453cc9b6c441eae10dd1481ff673a9cb923ad"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.009559 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.012976 4837 generic.go:334] "Generic (PLEG): container finished" podID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerID="9d0c63a56b505c323247613cb05d39bfe8227df3c32c1ff6ebcabfa1e90233e6" exitCode=0 Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.013049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerDied","Data":"9d0c63a56b505c323247613cb05d39bfe8227df3c32c1ff6ebcabfa1e90233e6"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.013087 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66dc5bc7b8-ttchd" event={"ID":"06af3d93-c582-4d5c-97d0-6a0250b1eb10","Type":"ContainerDied","Data":"28ca7e3ea0beaee25f8c27512f8e60012309156a94215dabf219cd2daca357c5"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.013125 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ca7e3ea0beaee25f8c27512f8e60012309156a94215dabf219cd2daca357c5" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.015044 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23126d10-0ad7-45e9-94e3-3e34116c20d8","Type":"ContainerDied","Data":"c28dbbd5d84b3f270ac6edb06187328bd34807ae9cf8ba2af98e74e863739a57"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.015192 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.021756 4837 generic.go:334] "Generic (PLEG): container finished" podID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerID="f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b" exitCode=0 Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.021786 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerDied","Data":"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.025631 4837 generic.go:334] "Generic (PLEG): container finished" podID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerID="e8db386829c99e752736e8a3a56fdc0ddce5b550d7fd78fdcfb1d074bd8ca628" exitCode=0 Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.025683 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerDied","Data":"e8db386829c99e752736e8a3a56fdc0ddce5b550d7fd78fdcfb1d074bd8ca628"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.025726 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7d64002d-6a47-4928-9495-91a58f4b2a5c","Type":"ContainerDied","Data":"87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.025738 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87ebcac171131819db001b6c87b4dfacb2dcf5de3b54fb69ad8687c9f317e248" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.027615 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6315-account-delete-ddhmt" event={"ID":"595b1839-8525-4b91-ab30-fd800710cf31","Type":"ContainerDied","Data":"d8c190e33adde7954694055619eab50e0af752af97028a89bcf0b0312c3da7a1"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.027679 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6315-account-delete-ddhmt" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.030812 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic192-account-delete-qzdmf" event={"ID":"6873cb35-815f-44ae-b38d-bb7ceb806b9d","Type":"ContainerDied","Data":"0914ae8ca47f0a9342bbeb91f2ddb67b624fec7e9654b08e871f232252327823"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.030837 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0914ae8ca47f0a9342bbeb91f2ddb67b624fec7e9654b08e871f232252327823" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.033354 4837 scope.go:117] "RemoveContainer" containerID="60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.053401 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.053929 4837 generic.go:334] "Generic (PLEG): container finished" podID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerID="2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40" exitCode=0 Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.053982 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerDied","Data":"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.054006 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17575b0a-a27e-4cc5-a84a-0ed9adf0ed65","Type":"ContainerDied","Data":"8414978f862c1ae0fe709e1d354a9339ff84e7723f35ffc1823a5b043b3ae0bd"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.054074 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.056751 4837 generic.go:334] "Generic (PLEG): container finished" podID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" containerID="afc5fac4c4bdbcdfcb537ec4258dde192ccbd7e0b4e6a075117de4ead1c2405d" exitCode=0 Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.056834 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7f2490fe-759d-4263-ae63-0f69c4f78c7d","Type":"ContainerDied","Data":"afc5fac4c4bdbcdfcb537ec4258dde192ccbd7e0b4e6a075117de4ead1c2405d"} Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.056883 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.057113 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.059166 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123166 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9t4g\" (UniqueName: \"kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123198 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2brw\" (UniqueName: \"kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123232 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123247 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123265 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123293 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123312 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123333 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123372 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123396 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123425 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123439 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkz9s\" (UniqueName: \"kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s\") pod \"6873cb35-815f-44ae-b38d-bb7ceb806b9d\" (UID: \"6873cb35-815f-44ae-b38d-bb7ceb806b9d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123462 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123495 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123525 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123554 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs\") pod \"7d64002d-6a47-4928-9495-91a58f4b2a5c\" (UID: \"7d64002d-6a47-4928-9495-91a58f4b2a5c\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.123592 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data\") pod \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\" (UID: \"06af3d93-c582-4d5c-97d0-6a0250b1eb10\") " Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.123962 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.124004 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data podName:96e5feff-74ce-434a-8606-85c51844deb9 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:27.123989939 +0000 UTC m=+1423.965597394 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data") pod "rabbitmq-cell1-server-0" (UID: "96e5feff-74ce-434a-8606-85c51844deb9") : configmap "rabbitmq-cell1-config-data" not found Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.124232 4837 scope.go:117] "RemoveContainer" containerID="c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.139718 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g" (OuterVolumeSpecName: "kube-api-access-n9t4g") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "kube-api-access-n9t4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.145486 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw" (OuterVolumeSpecName: "kube-api-access-p2brw") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "kube-api-access-p2brw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.146016 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.146144 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs" (OuterVolumeSpecName: "logs") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.147044 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223\": container with ID starting with c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223 not found: ID does not exist" containerID="c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.147113 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223"} err="failed to get container status \"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223\": rpc error: code = NotFound desc = could not find container \"c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223\": container with ID starting with c9e902a909a987019f97ad5b5fb890180cae6e58d3aff3b081099df5e4292223 not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.147143 4837 scope.go:117] "RemoveContainer" containerID="60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.147663 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs" (OuterVolumeSpecName: "logs") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.158882 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500\": container with ID starting with 60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500 not found: ID does not exist" containerID="60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.160492 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500"} err="failed to get container status \"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500\": rpc error: code = NotFound desc = could not find container \"60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500\": container with ID starting with 60d63ce6e9692a3544408a6a955d26054f3c5f9a22f6893b8ffb8cef4b067500 not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.160594 4837 scope.go:117] "RemoveContainer" containerID="d82ff67647eae0dc78cadc5398b75190de9934067eddeef92311560e0c706fd1" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.159804 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.191896 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.194717 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.204554 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance6315-account-delete-ddhmt"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.205838 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s" (OuterVolumeSpecName: "kube-api-access-jkz9s") pod "6873cb35-815f-44ae-b38d-bb7ceb806b9d" (UID: "6873cb35-815f-44ae-b38d-bb7ceb806b9d"). InnerVolumeSpecName "kube-api-access-jkz9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.205873 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts" (OuterVolumeSpecName: "scripts") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225738 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9t4g\" (UniqueName: \"kubernetes.io/projected/7d64002d-6a47-4928-9495-91a58f4b2a5c-kube-api-access-n9t4g\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225767 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2brw\" (UniqueName: \"kubernetes.io/projected/06af3d93-c582-4d5c-97d0-6a0250b1eb10-kube-api-access-p2brw\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225778 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225789 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225801 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225810 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkz9s\" (UniqueName: \"kubernetes.io/projected/6873cb35-815f-44ae-b38d-bb7ceb806b9d-kube-api-access-jkz9s\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225818 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d64002d-6a47-4928-9495-91a58f4b2a5c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225831 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06af3d93-c582-4d5c-97d0-6a0250b1eb10-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.225838 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d64002d-6a47-4928-9495-91a58f4b2a5c-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.238558 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.247435 4837 scope.go:117] "RemoveContainer" containerID="f2e57a498d2d7002b6196b4a7d582862dc279abc76d6c4f29681768a4d3c6a6f" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.272141 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.275551 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.280503 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.287063 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data" (OuterVolumeSpecName: "config-data") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.299943 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.301741 4837 scope.go:117] "RemoveContainer" containerID="0c61265ca20b5940cf1844ccd442b04f5de4c5d17b8de5b7bfbba2dd3bbb9d3c" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.327027 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.330277 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.330311 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.330324 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.330394 4837 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.330454 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data podName:0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087 nodeName:}" failed. No retries permitted until 2025-10-01 07:29:27.330436476 +0000 UTC m=+1424.172043931 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data") pod "rabbitmq-server-0" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087") : configmap "rabbitmq-config-data" not found Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.335302 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.335974 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.344801 4837 scope.go:117] "RemoveContainer" containerID="5f3d01a41efebb2a65c4e6e279d47ec4c2349d13a91fa0e7c788a05bf6c23dbd" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.350266 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.357875 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.364455 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.366881 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7d64002d-6a47-4928-9495-91a58f4b2a5c" (UID: "7d64002d-6a47-4928-9495-91a58f4b2a5c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.368097 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.373846 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.377157 4837 scope.go:117] "RemoveContainer" containerID="0526bde28972f0419714df8a5e921f83607729d39715b762ffd0a2c5dd4a7841" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.380594 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.387141 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.390026 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data" (OuterVolumeSpecName: "config-data") pod "06af3d93-c582-4d5c-97d0-6a0250b1eb10" (UID: "06af3d93-c582-4d5c-97d0-6a0250b1eb10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.392130 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.398007 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.398576 4837 scope.go:117] "RemoveContainer" containerID="ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.404762 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.409359 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.413532 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.415588 4837 scope.go:117] "RemoveContainer" containerID="ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.416040 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b\": container with ID starting with ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b not found: ID does not exist" containerID="ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.416077 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b"} err="failed to get container status \"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b\": rpc error: code = NotFound desc = could not find container \"ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b\": container with ID starting with ada2006f75f4546007c620a9b53b534a93a51bb31100e719ddede107f807922b not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.416169 4837 scope.go:117] "RemoveContainer" containerID="0e91c05a507a757b672ef0ff4aa99493b289fd25a5e7eb5cf9c6c892390a95bf" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.417632 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.421905 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.430798 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs\") pod \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.430904 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle\") pod \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431210 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config\") pod \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431238 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data\") pod \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431265 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz82z\" (UniqueName: \"kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z\") pod \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\" (UID: \"7f2490fe-759d-4263-ae63-0f69c4f78c7d\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431656 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431675 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431683 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431749 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d64002d-6a47-4928-9495-91a58f4b2a5c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431758 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af3d93-c582-4d5c-97d0-6a0250b1eb10-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.431930 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data" (OuterVolumeSpecName: "config-data") pod "7f2490fe-759d-4263-ae63-0f69c4f78c7d" (UID: "7f2490fe-759d-4263-ae63-0f69c4f78c7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.432105 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7f2490fe-759d-4263-ae63-0f69c4f78c7d" (UID: "7f2490fe-759d-4263-ae63-0f69c4f78c7d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.434339 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z" (OuterVolumeSpecName: "kube-api-access-lz82z") pod "7f2490fe-759d-4263-ae63-0f69c4f78c7d" (UID: "7f2490fe-759d-4263-ae63-0f69c4f78c7d"). InnerVolumeSpecName "kube-api-access-lz82z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.450523 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f2490fe-759d-4263-ae63-0f69c4f78c7d" (UID: "7f2490fe-759d-4263-ae63-0f69c4f78c7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.464506 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "7f2490fe-759d-4263-ae63-0f69c4f78c7d" (UID: "7f2490fe-759d-4263-ae63-0f69c4f78c7d"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.497371 4837 scope.go:117] "RemoveContainer" containerID="121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.516961 4837 scope.go:117] "RemoveContainer" containerID="d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.533186 4837 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.533219 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2490fe-759d-4263-ae63-0f69c4f78c7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.533248 4837 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.533261 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f2490fe-759d-4263-ae63-0f69c4f78c7d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.533274 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz82z\" (UniqueName: \"kubernetes.io/projected/7f2490fe-759d-4263-ae63-0f69c4f78c7d-kube-api-access-lz82z\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.559052 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.567098 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.568532 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.568572 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" containerName="nova-cell0-conductor-conductor" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.627518 4837 scope.go:117] "RemoveContainer" containerID="121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.627874 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191\": container with ID starting with 121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191 not found: ID does not exist" containerID="121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.627918 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191"} err="failed to get container status \"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191\": rpc error: code = NotFound desc = could not find container \"121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191\": container with ID starting with 121413c3d466383b875f2c476ecce9cdb5eadf43ac4d44a049a5038396e7e191 not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.627968 4837 scope.go:117] "RemoveContainer" containerID="d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.628221 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a\": container with ID starting with d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a not found: ID does not exist" containerID="d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.628256 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a"} err="failed to get container status \"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a\": rpc error: code = NotFound desc = could not find container \"d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a\": container with ID starting with d23c4e997d1c82af11a84ed466544fc9ed0afc013b313c87c94e0e106563124a not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.628281 4837 scope.go:117] "RemoveContainer" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.680170 4837 scope.go:117] "RemoveContainer" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" Oct 01 07:29:19 crc kubenswrapper[4837]: E1001 07:29:19.680644 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088\": container with ID starting with 2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088 not found: ID does not exist" containerID="2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.680677 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088"} err="failed to get container status \"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088\": rpc error: code = NotFound desc = could not find container \"2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088\": container with ID starting with 2ff8a4761c24128255f9f4d28cbc3ba3e68153a63fbbb839fbe6a0bba227e088 not found: ID does not exist" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.680776 4837 scope.go:117] "RemoveContainer" containerID="470c5f7992ab6223b3d21e3422fa5783032efa7c901479e728e382c4209680ac" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.710407 4837 scope.go:117] "RemoveContainer" containerID="b1a25a5631787393515c8e82608962228bfb05a6623f191895b95414994acc1b" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.736607 4837 scope.go:117] "RemoveContainer" containerID="e4a74db66f683da3b67436f59d8f19271ded325f5539b6a7b61f313bce201a1c" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.771312 4837 scope.go:117] "RemoveContainer" containerID="2422f422ed0b438a27d2a54ccb39f04357a6213eac4f80a06db39e4b2ebad2be" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.826015 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" path="/var/lib/kubelet/pods/08552fbf-b050-46ec-b8bf-83cde04f87e6/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.827602 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" path="/var/lib/kubelet/pods/17575b0a-a27e-4cc5-a84a-0ed9adf0ed65/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.829372 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" path="/var/lib/kubelet/pods/23126d10-0ad7-45e9-94e3-3e34116c20d8/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.830002 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" path="/var/lib/kubelet/pods/2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.830531 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="595b1839-8525-4b91-ab30-fd800710cf31" path="/var/lib/kubelet/pods/595b1839-8525-4b91-ab30-fd800710cf31/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.831564 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a3bb280-8572-448d-a3af-4c33993fc232" path="/var/lib/kubelet/pods/5a3bb280-8572-448d-a3af-4c33993fc232/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.832846 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6063ad95-cbaf-4798-899f-195d4428f246" path="/var/lib/kubelet/pods/6063ad95-cbaf-4798-899f-195d4428f246/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.833670 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" path="/var/lib/kubelet/pods/6212cb23-af83-4bcf-8328-91b14a62844b/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.835851 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" path="/var/lib/kubelet/pods/73e65e7c-ced2-4e38-9841-3f383471b275/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.839421 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" path="/var/lib/kubelet/pods/7f15a72f-f4ec-4e07-ae6d-7eef5700f543/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.839974 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5e54c4-ad98-4f26-9681-189997f3bd64" path="/var/lib/kubelet/pods/aa5e54c4-ad98-4f26-9681-189997f3bd64/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.842339 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2851153-f304-4041-9de2-045c0d92526d" path="/var/lib/kubelet/pods/b2851153-f304-4041-9de2-045c0d92526d/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.843138 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" path="/var/lib/kubelet/pods/c46a0ffe-6c41-4d34-bbe9-8db505bbe46b/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.843967 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce4369c-e009-4ae1-bc07-65dff0005807" path="/var/lib/kubelet/pods/dce4369c-e009-4ae1-bc07-65dff0005807/volumes" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.858297 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.858730 4837 scope.go:117] "RemoveContainer" containerID="ea330836b8714a11092cc9a2f567f8f4866b7798c3ad93cc78b29650a25a1901" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.882655 4837 scope.go:117] "RemoveContainer" containerID="a708b9492c517db3c4ac65ac6d8e06a0899c2e2382102fdd99be860d6cabafbe" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.911747 4837 scope.go:117] "RemoveContainer" containerID="4db0657ecb1d6bab9cfa0aca6d6c4da781d4dc1788c9788ff50b8ac94c614f0d" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941138 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941193 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941285 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941325 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941362 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941428 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g85kd\" (UniqueName: \"kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941451 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941466 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941482 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.941800 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.942896 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.943200 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.943215 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.949836 4837 scope.go:117] "RemoveContainer" containerID="942fd5c680f88e7c4fd1e2e59bf8058b4e347d103f7e8b089ff68968743b0e69" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.950060 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd" (OuterVolumeSpecName: "kube-api-access-g85kd") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "kube-api-access-g85kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:19 crc kubenswrapper[4837]: I1001 07:29:19.958843 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets" (OuterVolumeSpecName: "secrets") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:19.968772 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.003924 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle podName:7dcde73a-d928-45af-914d-3b1f5f7f899b nodeName:}" failed. No retries permitted until 2025-10-01 07:29:20.503864059 +0000 UTC m=+1417.345471514 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b") : error deleting /var/lib/kubelet/pods/7dcde73a-d928-45af-914d-3b1f5f7f899b/volume-subpaths: remove /var/lib/kubelet/pods/7dcde73a-d928-45af-914d-3b1f5f7f899b/volume-subpaths: no such file or directory Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.008956 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043910 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043943 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043956 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7dcde73a-d928-45af-914d-3b1f5f7f899b-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043967 4837 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043979 4837 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.043990 4837 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.044003 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g85kd\" (UniqueName: \"kubernetes.io/projected/7dcde73a-d928-45af-914d-3b1f5f7f899b-kube-api-access-g85kd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.044018 4837 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dcde73a-d928-45af-914d-3b1f5f7f899b-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.061942 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.071076 4837 generic.go:334] "Generic (PLEG): container finished" podID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" exitCode=0 Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.071140 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerDied","Data":"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.071163 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7dcde73a-d928-45af-914d-3b1f5f7f899b","Type":"ContainerDied","Data":"6e4114f7e83256f7322cb3d583e2616219837e99d678ca299638e4b02d3305aa"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.071215 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.074349 4837 generic.go:334] "Generic (PLEG): container finished" podID="96e5feff-74ce-434a-8606-85c51844deb9" containerID="3f3ea1fd9b8bc426d4733c4c8e9a603415784a810c094ebe41057d96f5b75c2e" exitCode=0 Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.074401 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerDied","Data":"3f3ea1fd9b8bc426d4733c4c8e9a603415784a810c094ebe41057d96f5b75c2e"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.076177 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7f2490fe-759d-4263-ae63-0f69c4f78c7d","Type":"ContainerDied","Data":"3c99a37f87139fe330d27863ab221dcb31ac8a65601421c6d91492703b1831c6"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.076232 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.083664 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerStarted","Data":"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.088936 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4140f175-0421-48d7-84e1-11fd6cdcea52/ovn-northd/0.log" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.089056 4837 generic.go:334] "Generic (PLEG): container finished" podID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" exitCode=139 Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.089110 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerDied","Data":"648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23"} Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.100899 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic192-account-delete-qzdmf" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.101759 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66dc5bc7b8-ttchd" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.101852 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.108349 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xg94g" podStartSLOduration=5.031630825 podStartE2EDuration="9.108333073s" podCreationTimestamp="2025-10-01 07:29:11 +0000 UTC" firstStartedPulling="2025-10-01 07:29:15.550759324 +0000 UTC m=+1412.392366779" lastFinishedPulling="2025-10-01 07:29:19.627461572 +0000 UTC m=+1416.469069027" observedRunningTime="2025-10-01 07:29:20.104715804 +0000 UTC m=+1416.946323279" watchObservedRunningTime="2025-10-01 07:29:20.108333073 +0000 UTC m=+1416.949940528" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.139430 4837 scope.go:117] "RemoveContainer" containerID="6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.144171 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.145408 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.154402 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapic192-account-delete-qzdmf"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.163856 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.167831 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.170643 4837 scope.go:117] "RemoveContainer" containerID="50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.187420 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.209928 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.211551 4837 scope.go:117] "RemoveContainer" containerID="2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.217922 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.222905 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-66dc5bc7b8-ttchd"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.234143 4837 scope.go:117] "RemoveContainer" containerID="ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.283026 4837 scope.go:117] "RemoveContainer" containerID="6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.283498 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c\": container with ID starting with 6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c not found: ID does not exist" containerID="6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.283525 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c"} err="failed to get container status \"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c\": rpc error: code = NotFound desc = could not find container \"6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c\": container with ID starting with 6f6718af40a944e3cc1723689d60bb47e6117ec7604aee09f6687604ae3e463c not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.283547 4837 scope.go:117] "RemoveContainer" containerID="50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.284086 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a\": container with ID starting with 50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a not found: ID does not exist" containerID="50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.284103 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a"} err="failed to get container status \"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a\": rpc error: code = NotFound desc = could not find container \"50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a\": container with ID starting with 50e40fd61019e1243f8606ceacd5e4c6a3deb9f1eef26c29819cea1e6cc8e00a not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.284115 4837 scope.go:117] "RemoveContainer" containerID="2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.289180 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40\": container with ID starting with 2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40 not found: ID does not exist" containerID="2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.289209 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40"} err="failed to get container status \"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40\": rpc error: code = NotFound desc = could not find container \"2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40\": container with ID starting with 2eea24b2acf20407ebe8d425603dbbdde7fd6d1f9d09ac877a2f8db9532c7f40 not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.289224 4837 scope.go:117] "RemoveContainer" containerID="ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.289438 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db\": container with ID starting with ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db not found: ID does not exist" containerID="ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.289459 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db"} err="failed to get container status \"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db\": rpc error: code = NotFound desc = could not find container \"ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db\": container with ID starting with ce8bd808610167d21bc5eb3c662172e9021e9a7351300fe3c6439a3767e991db not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.289471 4837 scope.go:117] "RemoveContainer" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.320245 4837 scope.go:117] "RemoveContainer" containerID="10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.454434 4837 scope.go:117] "RemoveContainer" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.454921 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e\": container with ID starting with 2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e not found: ID does not exist" containerID="2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.454962 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e"} err="failed to get container status \"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e\": rpc error: code = NotFound desc = could not find container \"2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e\": container with ID starting with 2ae089db63dbb866ec38c4538ac2d393cef36d7a64684436417c0045fcb6421e not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.454988 4837 scope.go:117] "RemoveContainer" containerID="10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.457076 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea\": container with ID starting with 10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea not found: ID does not exist" containerID="10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.457102 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea"} err="failed to get container status \"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea\": rpc error: code = NotFound desc = could not find container \"10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea\": container with ID starting with 10914286f4e0b74947fbd02ada61c89d9f97379d495b7723162aa791ff6c12ea not found: ID does not exist" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.457123 4837 scope.go:117] "RemoveContainer" containerID="afc5fac4c4bdbcdfcb537ec4258dde192ccbd7e0b4e6a075117de4ead1c2405d" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.474230 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4140f175-0421-48d7-84e1-11fd6cdcea52/ovn-northd/0.log" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.474299 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550231 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550819 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550857 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550881 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550920 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550950 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpzrw\" (UniqueName: \"kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.550987 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts\") pod \"4140f175-0421-48d7-84e1-11fd6cdcea52\" (UID: \"4140f175-0421-48d7-84e1-11fd6cdcea52\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.551115 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") pod \"7dcde73a-d928-45af-914d-3b1f5f7f899b\" (UID: \"7dcde73a-d928-45af-914d-3b1f5f7f899b\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.551741 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.551921 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts" (OuterVolumeSpecName: "scripts") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.551959 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config" (OuterVolumeSpecName: "config") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.563684 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw" (OuterVolumeSpecName: "kube-api-access-dpzrw") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "kube-api-access-dpzrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.566110 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dcde73a-d928-45af-914d-3b1f5f7f899b" (UID: "7dcde73a-d928-45af-914d-3b1f5f7f899b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.579032 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.632912 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.633488 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.633924 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.633954 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.635634 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.637769 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.645917 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.647197 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:20 crc kubenswrapper[4837]: E1001 07:29:20.647249 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653507 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpzrw\" (UniqueName: \"kubernetes.io/projected/4140f175-0421-48d7-84e1-11fd6cdcea52-kube-api-access-dpzrw\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653548 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653562 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dcde73a-d928-45af-914d-3b1f5f7f899b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653574 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653586 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4140f175-0421-48d7-84e1-11fd6cdcea52-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653598 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.653610 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4140f175-0421-48d7-84e1-11fd6cdcea52-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.662938 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4140f175-0421-48d7-84e1-11fd6cdcea52" (UID: "4140f175-0421-48d7-84e1-11fd6cdcea52"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.688915 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.726070 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.730743 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.737091 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.752180 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.756145 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxdj2\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.756345 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.757953 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758265 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758311 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758387 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758422 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758457 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758512 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758539 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.758574 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data\") pod \"96e5feff-74ce-434a-8606-85c51844deb9\" (UID: \"96e5feff-74ce-434a-8606-85c51844deb9\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.760958 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.761162 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.761996 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.763601 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.764909 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.765404 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.765790 4837 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4140f175-0421-48d7-84e1-11fd6cdcea52-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.765865 4837 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.765918 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.766086 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.766451 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.774112 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.774727 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info" (OuterVolumeSpecName: "pod-info") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.798151 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.805088 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2" (OuterVolumeSpecName: "kube-api-access-vxdj2") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "kube-api-access-vxdj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.812012 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf" (OuterVolumeSpecName: "server-conf") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.827249 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data" (OuterVolumeSpecName: "config-data") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.855878 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "96e5feff-74ce-434a-8606-85c51844deb9" (UID: "96e5feff-74ce-434a-8606-85c51844deb9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867275 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867324 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867354 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867370 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867384 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867431 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867452 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgd6l\" (UniqueName: \"kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867475 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867489 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867513 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867529 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867545 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867562 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867585 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5vz4\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867601 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867643 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867718 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf\") pod \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\" (UID: \"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867734 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.867749 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs\") pod \"943a976f-044a-4b42-b08c-baeae7b85f42\" (UID: \"943a976f-044a-4b42-b08c-baeae7b85f42\") " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868036 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868053 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868062 4837 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868070 4837 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/96e5feff-74ce-434a-8606-85c51844deb9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868079 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/96e5feff-74ce-434a-8606-85c51844deb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868087 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxdj2\" (UniqueName: \"kubernetes.io/projected/96e5feff-74ce-434a-8606-85c51844deb9-kube-api-access-vxdj2\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.868098 4837 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/96e5feff-74ce-434a-8606-85c51844deb9-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.869105 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.869403 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.869545 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.869898 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.873290 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info" (OuterVolumeSpecName: "pod-info") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.873372 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4" (OuterVolumeSpecName: "kube-api-access-j5vz4") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "kube-api-access-j5vz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.873304 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.873805 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts" (OuterVolumeSpecName: "scripts") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.874830 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.874957 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l" (OuterVolumeSpecName: "kube-api-access-wgd6l") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "kube-api-access-wgd6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.882656 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.886777 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.892303 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data" (OuterVolumeSpecName: "config-data") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.892621 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data" (OuterVolumeSpecName: "config-data") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.905146 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.910454 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf" (OuterVolumeSpecName: "server-conf") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.915271 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.923286 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "943a976f-044a-4b42-b08c-baeae7b85f42" (UID: "943a976f-044a-4b42-b08c-baeae7b85f42"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.957689 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" (UID: "0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969573 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969604 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969619 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969630 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969644 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5vz4\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-kube-api-access-j5vz4\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969655 4837 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969664 4837 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969674 4837 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969683 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969710 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969721 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969732 4837 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969766 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969781 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969793 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969803 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969813 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgd6l\" (UniqueName: \"kubernetes.io/projected/943a976f-044a-4b42-b08c-baeae7b85f42-kube-api-access-wgd6l\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969824 4837 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.969834 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/943a976f-044a-4b42-b08c-baeae7b85f42-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:20 crc kubenswrapper[4837]: I1001 07:29:20.985133 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.071541 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.114515 4837 generic.go:334] "Generic (PLEG): container finished" podID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerID="afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f" exitCode=0 Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.114597 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.114623 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerDied","Data":"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.114668 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087","Type":"ContainerDied","Data":"dc29fcfd5af62b329b544cd0c06967919cdfbcde1b023e88bda0c6c9ccecb870"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.114729 4837 scope.go:117] "RemoveContainer" containerID="afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.116753 4837 generic.go:334] "Generic (PLEG): container finished" podID="943a976f-044a-4b42-b08c-baeae7b85f42" containerID="6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3" exitCode=0 Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.116970 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5799b746f7-mzft5" event={"ID":"943a976f-044a-4b42-b08c-baeae7b85f42","Type":"ContainerDied","Data":"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.117077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5799b746f7-mzft5" event={"ID":"943a976f-044a-4b42-b08c-baeae7b85f42","Type":"ContainerDied","Data":"700b94b97abe1787c669a876fca142041a7d8e45ca92e097e477d59f24bd52a0"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.116978 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5799b746f7-mzft5" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.121127 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.121673 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"96e5feff-74ce-434a-8606-85c51844deb9","Type":"ContainerDied","Data":"eb4b6ed36fb446c04f97e906a0e9933f49802e8c4266043a294f9ccd3aeeb126"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.130073 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4140f175-0421-48d7-84e1-11fd6cdcea52/ovn-northd/0.log" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.130298 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4140f175-0421-48d7-84e1-11fd6cdcea52","Type":"ContainerDied","Data":"673e7ff209cad99253c12a17b7d404dfd9f4f5912b7718e5f90531dd3bf3ef2a"} Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.130415 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.139314 4837 scope.go:117] "RemoveContainer" containerID="69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.165652 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.174268 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5799b746f7-mzft5"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.180993 4837 scope.go:117] "RemoveContainer" containerID="afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f" Oct 01 07:29:21 crc kubenswrapper[4837]: E1001 07:29:21.182000 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f\": container with ID starting with afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f not found: ID does not exist" containerID="afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.182038 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f"} err="failed to get container status \"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f\": rpc error: code = NotFound desc = could not find container \"afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f\": container with ID starting with afeaac802615772dd63cc743558a770291b1de6033cbfb9e3901170065ab105f not found: ID does not exist" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.182065 4837 scope.go:117] "RemoveContainer" containerID="69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139" Oct 01 07:29:21 crc kubenswrapper[4837]: E1001 07:29:21.183002 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139\": container with ID starting with 69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139 not found: ID does not exist" containerID="69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.183045 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139"} err="failed to get container status \"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139\": rpc error: code = NotFound desc = could not find container \"69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139\": container with ID starting with 69728e2e415086362fc974593e9f984319f33bd8385d936e6275d5af1f773139 not found: ID does not exist" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.183071 4837 scope.go:117] "RemoveContainer" containerID="6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.190790 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.202828 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.202887 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.212050 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.222747 4837 scope.go:117] "RemoveContainer" containerID="6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3" Oct 01 07:29:21 crc kubenswrapper[4837]: E1001 07:29:21.223327 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3\": container with ID starting with 6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3 not found: ID does not exist" containerID="6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.223370 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3"} err="failed to get container status \"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3\": rpc error: code = NotFound desc = could not find container \"6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3\": container with ID starting with 6967167f62db4e503488a01d06a5a2b939d7538d3d3406eb62a7fe382a86f9e3 not found: ID does not exist" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.223409 4837 scope.go:117] "RemoveContainer" containerID="3f3ea1fd9b8bc426d4733c4c8e9a603415784a810c094ebe41057d96f5b75c2e" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.232278 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.239209 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.245411 4837 scope.go:117] "RemoveContainer" containerID="101c60691fb7c41b74bbf3f7d9a52d44e7d583f292d0df8ca0ed6767687a2c81" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.266756 4837 scope.go:117] "RemoveContainer" containerID="a699f84318a8fa1ea08dd47672464162bcd065e0a5b23d594adac792bb355f6b" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.286162 4837 scope.go:117] "RemoveContainer" containerID="648f9ed191aed48b566a682412c5b2e8516bc03c4be4c11032fe671271fa6b23" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.827017 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" path="/var/lib/kubelet/pods/06af3d93-c582-4d5c-97d0-6a0250b1eb10/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.827902 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" path="/var/lib/kubelet/pods/0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.828958 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" path="/var/lib/kubelet/pods/4140f175-0421-48d7-84e1-11fd6cdcea52/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.829580 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6873cb35-815f-44ae-b38d-bb7ceb806b9d" path="/var/lib/kubelet/pods/6873cb35-815f-44ae-b38d-bb7ceb806b9d/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.830244 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" path="/var/lib/kubelet/pods/7d64002d-6a47-4928-9495-91a58f4b2a5c/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.831571 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" path="/var/lib/kubelet/pods/7dcde73a-d928-45af-914d-3b1f5f7f899b/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.832303 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" path="/var/lib/kubelet/pods/7f2490fe-759d-4263-ae63-0f69c4f78c7d/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.832889 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="943a976f-044a-4b42-b08c-baeae7b85f42" path="/var/lib/kubelet/pods/943a976f-044a-4b42-b08c-baeae7b85f42/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.834598 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e5feff-74ce-434a-8606-85c51844deb9" path="/var/lib/kubelet/pods/96e5feff-74ce-434a-8606-85c51844deb9/volumes" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.859538 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.889259 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.889518 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.985234 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data\") pod \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.985313 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs\") pod \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.985381 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle\") pod \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.985466 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9fhc\" (UniqueName: \"kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc\") pod \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.985528 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom\") pod \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\" (UID: \"9f871dbb-1e64-411b-acfa-39ed6b5b66ed\") " Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.986182 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs" (OuterVolumeSpecName: "logs") pod "9f871dbb-1e64-411b-acfa-39ed6b5b66ed" (UID: "9f871dbb-1e64-411b-acfa-39ed6b5b66ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.989725 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc" (OuterVolumeSpecName: "kube-api-access-b9fhc") pod "9f871dbb-1e64-411b-acfa-39ed6b5b66ed" (UID: "9f871dbb-1e64-411b-acfa-39ed6b5b66ed"). InnerVolumeSpecName "kube-api-access-b9fhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:21 crc kubenswrapper[4837]: I1001 07:29:21.998854 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9f871dbb-1e64-411b-acfa-39ed6b5b66ed" (UID: "9f871dbb-1e64-411b-acfa-39ed6b5b66ed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.007699 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f871dbb-1e64-411b-acfa-39ed6b5b66ed" (UID: "9f871dbb-1e64-411b-acfa-39ed6b5b66ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.025458 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data" (OuterVolumeSpecName: "config-data") pod "9f871dbb-1e64-411b-acfa-39ed6b5b66ed" (UID: "9f871dbb-1e64-411b-acfa-39ed6b5b66ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.037272 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.047655 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.086910 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htgkt\" (UniqueName: \"kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt\") pod \"9c09c611-b859-4650-b4b0-96738b65fc45\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.086966 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data\") pod \"9c09c611-b859-4650-b4b0-96738b65fc45\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.086989 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle\") pod \"9c09c611-b859-4650-b4b0-96738b65fc45\" (UID: \"9c09c611-b859-4650-b4b0-96738b65fc45\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.087907 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.087927 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.087938 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.087947 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9fhc\" (UniqueName: \"kubernetes.io/projected/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-kube-api-access-b9fhc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.087956 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f871dbb-1e64-411b-acfa-39ed6b5b66ed-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.090863 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt" (OuterVolumeSpecName: "kube-api-access-htgkt") pod "9c09c611-b859-4650-b4b0-96738b65fc45" (UID: "9c09c611-b859-4650-b4b0-96738b65fc45"). InnerVolumeSpecName "kube-api-access-htgkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.106403 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c09c611-b859-4650-b4b0-96738b65fc45" (UID: "9c09c611-b859-4650-b4b0-96738b65fc45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.107018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data" (OuterVolumeSpecName: "config-data") pod "9c09c611-b859-4650-b4b0-96738b65fc45" (UID: "9c09c611-b859-4650-b4b0-96738b65fc45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.153231 4837 generic.go:334] "Generic (PLEG): container finished" podID="9c09c611-b859-4650-b4b0-96738b65fc45" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" exitCode=0 Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.153267 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.153285 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c09c611-b859-4650-b4b0-96738b65fc45","Type":"ContainerDied","Data":"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.153550 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c09c611-b859-4650-b4b0-96738b65fc45","Type":"ContainerDied","Data":"acfb24c0dc96b2ced53b34659ef44aeb135e57d07544dc84f4597c41be3f48d9"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.153568 4837 scope.go:117] "RemoveContainer" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.155978 4837 generic.go:334] "Generic (PLEG): container finished" podID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerID="e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164" exitCode=0 Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.156246 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.156205 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerDied","Data":"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.156382 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-849c79f7db-h5jwx" event={"ID":"9f871dbb-1e64-411b-acfa-39ed6b5b66ed","Type":"ContainerDied","Data":"ab36d941091a7e924c9f3355720ded0ef841331ab576500130be7e7286ab52c5"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.159056 4837 generic.go:334] "Generic (PLEG): container finished" podID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerID="3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41" exitCode=0 Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.159137 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerDied","Data":"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.159153 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b46954cf-msjzx" event={"ID":"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da","Type":"ContainerDied","Data":"eab354248b04ca89f90412b44f455ea4155cdc7962efcd3b71b4bfdfe58cbb67"} Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.159167 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b46954cf-msjzx" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189041 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom\") pod \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189163 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data\") pod \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189318 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmckq\" (UniqueName: \"kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq\") pod \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189665 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle\") pod \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189755 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs\") pod \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\" (UID: \"8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da\") " Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.189821 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.190445 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htgkt\" (UniqueName: \"kubernetes.io/projected/9c09c611-b859-4650-b4b0-96738b65fc45-kube-api-access-htgkt\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.190470 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.190483 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c09c611-b859-4650-b4b0-96738b65fc45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.190573 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs" (OuterVolumeSpecName: "logs") pod "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" (UID: "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.192446 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" (UID: "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.192838 4837 scope.go:117] "RemoveContainer" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" Oct 01 07:29:22 crc kubenswrapper[4837]: E1001 07:29:22.193262 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8\": container with ID starting with 5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8 not found: ID does not exist" containerID="5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.193289 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8"} err="failed to get container status \"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8\": rpc error: code = NotFound desc = could not find container \"5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8\": container with ID starting with 5dbe6fc37fbfc6410e0890f200c984e84e2bbdc36da1308fda42f75643d4c7b8 not found: ID does not exist" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.193309 4837 scope.go:117] "RemoveContainer" containerID="e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.193590 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq" (OuterVolumeSpecName: "kube-api-access-lmckq") pod "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" (UID: "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da"). InnerVolumeSpecName "kube-api-access-lmckq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.195972 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.208356 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.215994 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" (UID: "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.216306 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-849c79f7db-h5jwx"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.219546 4837 scope.go:117] "RemoveContainer" containerID="44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.236752 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data" (OuterVolumeSpecName: "config-data") pod "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" (UID: "8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.238025 4837 scope.go:117] "RemoveContainer" containerID="e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164" Oct 01 07:29:22 crc kubenswrapper[4837]: E1001 07:29:22.238412 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164\": container with ID starting with e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164 not found: ID does not exist" containerID="e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.238472 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164"} err="failed to get container status \"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164\": rpc error: code = NotFound desc = could not find container \"e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164\": container with ID starting with e4c17bca3a7af5f08d54e59da2dcf9c06620c49e7a28d45e1354f1a8fa292164 not found: ID does not exist" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.238501 4837 scope.go:117] "RemoveContainer" containerID="44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666" Oct 01 07:29:22 crc kubenswrapper[4837]: E1001 07:29:22.238831 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666\": container with ID starting with 44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666 not found: ID does not exist" containerID="44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.238857 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666"} err="failed to get container status \"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666\": rpc error: code = NotFound desc = could not find container \"44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666\": container with ID starting with 44669bc766a9def5bd8be90fccbfad4246b2ecd278430562dd9663a053a46666 not found: ID does not exist" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.238896 4837 scope.go:117] "RemoveContainer" containerID="3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.264825 4837 scope.go:117] "RemoveContainer" containerID="4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.282464 4837 scope.go:117] "RemoveContainer" containerID="3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41" Oct 01 07:29:22 crc kubenswrapper[4837]: E1001 07:29:22.282996 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41\": container with ID starting with 3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41 not found: ID does not exist" containerID="3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.283040 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41"} err="failed to get container status \"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41\": rpc error: code = NotFound desc = could not find container \"3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41\": container with ID starting with 3947e3ba058afee6124c0216fc178fdd636d3be88b7b1f5a92b3f99833de1e41 not found: ID does not exist" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.283062 4837 scope.go:117] "RemoveContainer" containerID="4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144" Oct 01 07:29:22 crc kubenswrapper[4837]: E1001 07:29:22.283324 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144\": container with ID starting with 4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144 not found: ID does not exist" containerID="4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.283363 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144"} err="failed to get container status \"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144\": rpc error: code = NotFound desc = could not find container \"4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144\": container with ID starting with 4662076039be8c318056f61e94966ea6a7ed046b65ab48640eafe407302f3144 not found: ID does not exist" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.291576 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.291615 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-logs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.291628 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.291640 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.291652 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmckq\" (UniqueName: \"kubernetes.io/projected/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da-kube-api-access-lmckq\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.487260 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.493728 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-68b46954cf-msjzx"] Oct 01 07:29:22 crc kubenswrapper[4837]: I1001 07:29:22.938629 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xg94g" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="registry-server" probeResult="failure" output=< Oct 01 07:29:22 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 07:29:22 crc kubenswrapper[4837]: > Oct 01 07:29:23 crc kubenswrapper[4837]: I1001 07:29:23.834104 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" path="/var/lib/kubelet/pods/8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da/volumes" Oct 01 07:29:23 crc kubenswrapper[4837]: I1001 07:29:23.836005 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" path="/var/lib/kubelet/pods/9c09c611-b859-4650-b4b0-96738b65fc45/volumes" Oct 01 07:29:23 crc kubenswrapper[4837]: I1001 07:29:23.837447 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" path="/var/lib/kubelet/pods/9f871dbb-1e64-411b-acfa-39ed6b5b66ed/volumes" Oct 01 07:29:23 crc kubenswrapper[4837]: I1001 07:29:23.968777 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66dc5bc7b8-ttchd" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.164:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 07:29:23 crc kubenswrapper[4837]: I1001 07:29:23.968842 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66dc5bc7b8-ttchd" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.164:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.631611 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.632235 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.632633 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.632732 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.633878 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.637326 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.639236 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:25 crc kubenswrapper[4837]: E1001 07:29:25.639295 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:28 crc kubenswrapper[4837]: I1001 07:29:28.936837 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035391 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035439 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035467 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b46vh\" (UniqueName: \"kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035482 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035520 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035557 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.035647 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config\") pod \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\" (UID: \"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2\") " Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.046003 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.046144 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh" (OuterVolumeSpecName: "kube-api-access-b46vh") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "kube-api-access-b46vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.076595 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.084228 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.093598 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config" (OuterVolumeSpecName: "config") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.104032 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.111507 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" (UID: "d6e3523b-a91a-4d3d-85c7-b41e4006c0e2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137552 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137586 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-config\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137600 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137614 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b46vh\" (UniqueName: \"kubernetes.io/projected/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-kube-api-access-b46vh\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137627 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137639 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.137651 4837 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.252225 4837 generic.go:334] "Generic (PLEG): container finished" podID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerID="59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9" exitCode=0 Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.252279 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf464dc69-n54w7" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.252291 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerDied","Data":"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9"} Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.252343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf464dc69-n54w7" event={"ID":"d6e3523b-a91a-4d3d-85c7-b41e4006c0e2","Type":"ContainerDied","Data":"2152ae928c2d6986b8654a591590735c5c17e4168ef13956b92d25ece300243a"} Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.252371 4837 scope.go:117] "RemoveContainer" containerID="2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.291521 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.305435 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cf464dc69-n54w7"] Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.305912 4837 scope.go:117] "RemoveContainer" containerID="59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.341080 4837 scope.go:117] "RemoveContainer" containerID="2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0" Oct 01 07:29:29 crc kubenswrapper[4837]: E1001 07:29:29.341510 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0\": container with ID starting with 2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0 not found: ID does not exist" containerID="2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.341539 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0"} err="failed to get container status \"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0\": rpc error: code = NotFound desc = could not find container \"2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0\": container with ID starting with 2ca97cb66189725aae8d9551b5a06830adcc55ba8b9e5879c982d29fcdbbbdf0 not found: ID does not exist" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.341607 4837 scope.go:117] "RemoveContainer" containerID="59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9" Oct 01 07:29:29 crc kubenswrapper[4837]: E1001 07:29:29.341924 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9\": container with ID starting with 59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9 not found: ID does not exist" containerID="59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.341951 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9"} err="failed to get container status \"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9\": rpc error: code = NotFound desc = could not find container \"59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9\": container with ID starting with 59e286c0f1f5b2ecd28d4eec242bf79799ec438c2a724ae820134d664d87dad9 not found: ID does not exist" Oct 01 07:29:29 crc kubenswrapper[4837]: I1001 07:29:29.832489 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" path="/var/lib/kubelet/pods/d6e3523b-a91a-4d3d-85c7-b41e4006c0e2/volumes" Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.632510 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.633223 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.633534 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.633579 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.634550 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.636339 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.638335 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:30 crc kubenswrapper[4837]: E1001 07:29:30.638391 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:31 crc kubenswrapper[4837]: I1001 07:29:31.967911 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:32 crc kubenswrapper[4837]: I1001 07:29:32.042970 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:32 crc kubenswrapper[4837]: I1001 07:29:32.222393 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.305367 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xg94g" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="registry-server" containerID="cri-o://97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901" gracePeriod=2 Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.788773 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.830300 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pxbv\" (UniqueName: \"kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv\") pod \"e273fd30-371c-494e-99ce-6460b2a1dd00\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.830417 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities\") pod \"e273fd30-371c-494e-99ce-6460b2a1dd00\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.830607 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content\") pod \"e273fd30-371c-494e-99ce-6460b2a1dd00\" (UID: \"e273fd30-371c-494e-99ce-6460b2a1dd00\") " Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.831896 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities" (OuterVolumeSpecName: "utilities") pod "e273fd30-371c-494e-99ce-6460b2a1dd00" (UID: "e273fd30-371c-494e-99ce-6460b2a1dd00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.838895 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv" (OuterVolumeSpecName: "kube-api-access-2pxbv") pod "e273fd30-371c-494e-99ce-6460b2a1dd00" (UID: "e273fd30-371c-494e-99ce-6460b2a1dd00"). InnerVolumeSpecName "kube-api-access-2pxbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.933234 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pxbv\" (UniqueName: \"kubernetes.io/projected/e273fd30-371c-494e-99ce-6460b2a1dd00-kube-api-access-2pxbv\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.933288 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:33 crc kubenswrapper[4837]: I1001 07:29:33.971074 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e273fd30-371c-494e-99ce-6460b2a1dd00" (UID: "e273fd30-371c-494e-99ce-6460b2a1dd00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.034565 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e273fd30-371c-494e-99ce-6460b2a1dd00-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.321744 4837 generic.go:334] "Generic (PLEG): container finished" podID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerID="97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901" exitCode=0 Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.321799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerDied","Data":"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901"} Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.321831 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xg94g" event={"ID":"e273fd30-371c-494e-99ce-6460b2a1dd00","Type":"ContainerDied","Data":"758e9fc071b4e96ed478a1adbcc56981884cdf43ae091ea0e77b53eadd9334fa"} Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.321838 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xg94g" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.321853 4837 scope.go:117] "RemoveContainer" containerID="97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.356396 4837 scope.go:117] "RemoveContainer" containerID="f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.381907 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.394217 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xg94g"] Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.404309 4837 scope.go:117] "RemoveContainer" containerID="e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.433680 4837 scope.go:117] "RemoveContainer" containerID="97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901" Oct 01 07:29:34 crc kubenswrapper[4837]: E1001 07:29:34.434241 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901\": container with ID starting with 97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901 not found: ID does not exist" containerID="97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.434292 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901"} err="failed to get container status \"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901\": rpc error: code = NotFound desc = could not find container \"97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901\": container with ID starting with 97cf569d267cf4fdd927d4f4a825eb4a75ac0a49e5c89c144672a9ef26293901 not found: ID does not exist" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.434328 4837 scope.go:117] "RemoveContainer" containerID="f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b" Oct 01 07:29:34 crc kubenswrapper[4837]: E1001 07:29:34.434886 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b\": container with ID starting with f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b not found: ID does not exist" containerID="f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.434949 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b"} err="failed to get container status \"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b\": rpc error: code = NotFound desc = could not find container \"f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b\": container with ID starting with f38ed2e2e9361e42e6ccc7e67b636953827970aff75f9db2246297314bd03b8b not found: ID does not exist" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.434998 4837 scope.go:117] "RemoveContainer" containerID="e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb" Oct 01 07:29:34 crc kubenswrapper[4837]: E1001 07:29:34.435412 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb\": container with ID starting with e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb not found: ID does not exist" containerID="e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb" Oct 01 07:29:34 crc kubenswrapper[4837]: I1001 07:29:34.435488 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb"} err="failed to get container status \"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb\": rpc error: code = NotFound desc = could not find container \"e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb\": container with ID starting with e076c68749b778ca04f93cedb12bb5c83133309834085411021a9a0bcec364eb not found: ID does not exist" Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.632486 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.634097 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.634122 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.634908 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.634949 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.635837 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.637554 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:35 crc kubenswrapper[4837]: E1001 07:29:35.637582 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:35 crc kubenswrapper[4837]: I1001 07:29:35.826727 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" path="/var/lib/kubelet/pods/e273fd30-371c-494e-99ce-6460b2a1dd00/volumes" Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.632524 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.634252 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.634930 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.635039 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.635944 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.638962 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.641616 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 01 07:29:40 crc kubenswrapper[4837]: E1001 07:29:40.641750 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zpx7h" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.442640 4837 generic.go:334] "Generic (PLEG): container finished" podID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerID="4eb16d26f4116a6a74e9d0b4accc3991ac0da36440e5760b73b34784878eb561" exitCode=137 Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.442681 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"4eb16d26f4116a6a74e9d0b4accc3991ac0da36440e5760b73b34784878eb561"} Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.446097 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zpx7h_beb9ef6d-db8f-4747-b93c-fa927429a785/ovs-vswitchd/0.log" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.447175 4837 generic.go:334] "Generic (PLEG): container finished" podID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" exitCode=137 Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.447210 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerDied","Data":"64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900"} Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.832556 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zpx7h_beb9ef6d-db8f-4747-b93c-fa927429a785/ovs-vswitchd/0.log" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.833886 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.849403 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992483 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992572 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992661 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock\") pod \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992725 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992745 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992799 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") pod \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992828 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d76km\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km\") pod \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992877 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992902 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache\") pod \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\" (UID: \"a27ce0b1-38ff-4a33-be51-5ba651e8e075\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.992928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkcp4\" (UniqueName: \"kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4\") pod \"beb9ef6d-db8f-4747-b93c-fa927429a785\" (UID: \"beb9ef6d-db8f-4747-b93c-fa927429a785\") " Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.993393 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log" (OuterVolumeSpecName: "var-log") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.993493 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.993540 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run" (OuterVolumeSpecName: "var-run") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.993644 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib" (OuterVolumeSpecName: "var-lib") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.993961 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache" (OuterVolumeSpecName: "cache") pod "a27ce0b1-38ff-4a33-be51-5ba651e8e075" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.994201 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock" (OuterVolumeSpecName: "lock") pod "a27ce0b1-38ff-4a33-be51-5ba651e8e075" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:29:43 crc kubenswrapper[4837]: I1001 07:29:43.995408 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts" (OuterVolumeSpecName: "scripts") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.000170 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km" (OuterVolumeSpecName: "kube-api-access-d76km") pod "a27ce0b1-38ff-4a33-be51-5ba651e8e075" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075"). InnerVolumeSpecName "kube-api-access-d76km". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.000340 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a27ce0b1-38ff-4a33-be51-5ba651e8e075" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.000578 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "a27ce0b1-38ff-4a33-be51-5ba651e8e075" (UID: "a27ce0b1-38ff-4a33-be51-5ba651e8e075"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.000658 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4" (OuterVolumeSpecName: "kube-api-access-pkcp4") pod "beb9ef6d-db8f-4747-b93c-fa927429a785" (UID: "beb9ef6d-db8f-4747-b93c-fa927429a785"). InnerVolumeSpecName "kube-api-access-pkcp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094498 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094551 4837 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-log\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094569 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beb9ef6d-db8f-4747-b93c-fa927429a785-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094759 4837 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-lock\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094776 4837 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094792 4837 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094807 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094826 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d76km\" (UniqueName: \"kubernetes.io/projected/a27ce0b1-38ff-4a33-be51-5ba651e8e075-kube-api-access-d76km\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094845 4837 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/beb9ef6d-db8f-4747-b93c-fa927429a785-var-lib\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094861 4837 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a27ce0b1-38ff-4a33-be51-5ba651e8e075-cache\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.094878 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkcp4\" (UniqueName: \"kubernetes.io/projected/beb9ef6d-db8f-4747-b93c-fa927429a785-kube-api-access-pkcp4\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.123243 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.196526 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.486348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a27ce0b1-38ff-4a33-be51-5ba651e8e075","Type":"ContainerDied","Data":"9afbec9311de400244922af066bd3bc1dac95bb0d3f80c0b8798a6f61fca1cae"} Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.486470 4837 scope.go:117] "RemoveContainer" containerID="4eb16d26f4116a6a74e9d0b4accc3991ac0da36440e5760b73b34784878eb561" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.486767 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.499418 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zpx7h_beb9ef6d-db8f-4747-b93c-fa927429a785/ovs-vswitchd/0.log" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.500772 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zpx7h" event={"ID":"beb9ef6d-db8f-4747-b93c-fa927429a785","Type":"ContainerDied","Data":"cae936c3d8781371574591fd0f37891f1b579b4a93242b367bfd809455712aec"} Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.500970 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zpx7h" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.540324 4837 scope.go:117] "RemoveContainer" containerID="0bfa759f888d0212de7ee4b0712ad5d33add96370a9084085bce043b2d375d44" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.576685 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.586070 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.591887 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.596211 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-zpx7h"] Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.599269 4837 scope.go:117] "RemoveContainer" containerID="2067fe40fd9f9d75412077801cde319e8068868f481690bf6f201f36f8f71642" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.628410 4837 scope.go:117] "RemoveContainer" containerID="708ad62022e8e219ec2f8ff1ce8f2f1e128a8e14443aec6458580876f7636171" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.659147 4837 scope.go:117] "RemoveContainer" containerID="fd642928ddcba69b2aa30d3fdad7ae984f220ec2171638957d4a6f7678a20b45" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.691620 4837 scope.go:117] "RemoveContainer" containerID="06276e79f10f7789a399e46f0f9ba7ceccd3ca7926331f03e09147648c3966d4" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.720277 4837 scope.go:117] "RemoveContainer" containerID="80729d7b80fe7ca0a76deb60aded53df6ae5bafa464acedacb113c4fcfa46c02" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.741420 4837 scope.go:117] "RemoveContainer" containerID="b5626cefa6e75a0a68d30120dccd151dc40d2c9a0f69f966c98d8c027ef04148" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.766824 4837 scope.go:117] "RemoveContainer" containerID="dbb2c5bf1f5499683d8d50b2b442d96eef02c58748d98bc5851c45b9a40621df" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.789408 4837 scope.go:117] "RemoveContainer" containerID="feb371a16a28f5d9590fbbaf472765ac149570124d961c2b6244deaac26355e5" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.817970 4837 scope.go:117] "RemoveContainer" containerID="9290d642faa7786d6b897ad12a0754d4d61afe7b651852d4dab287495dddc287" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.841395 4837 scope.go:117] "RemoveContainer" containerID="2204c01a3c6692934f14277b4d5cbf6cca7a5c501820bb3fa3389ae18fea6f76" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.868565 4837 scope.go:117] "RemoveContainer" containerID="8caed28b92cf49fa4a285edcda33f9590fa05d856cd61e8bcce897cc1725a303" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.896440 4837 scope.go:117] "RemoveContainer" containerID="c67bf25539192ca5553d8c61e702cf3adae4424f5c481f14196a5218a1c14dde" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.942159 4837 scope.go:117] "RemoveContainer" containerID="6d2afe280ee355f22575011affdf80ede81294951985ed4fdfaf0a0a8bd04577" Oct 01 07:29:44 crc kubenswrapper[4837]: I1001 07:29:44.980475 4837 scope.go:117] "RemoveContainer" containerID="64af10eb19e015a4bb49f0476fa0de1957208eee2c24530167551546ab179900" Oct 01 07:29:45 crc kubenswrapper[4837]: I1001 07:29:45.014795 4837 scope.go:117] "RemoveContainer" containerID="d4c88e80869d08d3fba283e79c9626d15b3b003f89939cf6cd1d9bf6f597e99f" Oct 01 07:29:45 crc kubenswrapper[4837]: I1001 07:29:45.052521 4837 scope.go:117] "RemoveContainer" containerID="8d7fbe2134351cb2359c79efd5278c7c6f0897a13622bea6a551c459a08e0614" Oct 01 07:29:45 crc kubenswrapper[4837]: I1001 07:29:45.828489 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" path="/var/lib/kubelet/pods/a27ce0b1-38ff-4a33-be51-5ba651e8e075/volumes" Oct 01 07:29:45 crc kubenswrapper[4837]: I1001 07:29:45.833453 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" path="/var/lib/kubelet/pods/beb9ef6d-db8f-4747-b93c-fa927429a785/volumes" Oct 01 07:29:45 crc kubenswrapper[4837]: I1001 07:29:45.913812 4837 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod34a848f4-7b85-4463-bdae-f763e6f9e7e5"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod34a848f4-7b85-4463-bdae-f763e6f9e7e5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod34a848f4_7b85_4463_bdae_f763e6f9e7e5.slice" Oct 01 07:29:45 crc kubenswrapper[4837]: E1001 07:29:45.913887 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod34a848f4-7b85-4463-bdae-f763e6f9e7e5] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod34a848f4-7b85-4463-bdae-f763e6f9e7e5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod34a848f4_7b85_4463_bdae_f763e6f9e7e5.slice" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" Oct 01 07:29:46 crc kubenswrapper[4837]: I1001 07:29:46.526595 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df6b7c479-f24qr" Oct 01 07:29:46 crc kubenswrapper[4837]: I1001 07:29:46.554631 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:29:46 crc kubenswrapper[4837]: I1001 07:29:46.563837 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-df6b7c479-f24qr"] Oct 01 07:29:47 crc kubenswrapper[4837]: I1001 07:29:47.833848 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" path="/var/lib/kubelet/pods/34a848f4-7b85-4463-bdae-f763e6f9e7e5/volumes" Oct 01 07:29:53 crc kubenswrapper[4837]: I1001 07:29:53.080350 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:29:53 crc kubenswrapper[4837]: I1001 07:29:53.081292 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.164487 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj"] Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165805 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165830 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165859 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerName="nova-cell1-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165872 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerName="nova-cell1-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165892 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="dnsmasq-dns" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165907 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="dnsmasq-dns" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165931 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28035a28-28d4-46b3-b269-97a523cf731e" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165943 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="28035a28-28d4-46b3-b269-97a523cf731e" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165957 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165966 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.165980 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="mysql-bootstrap" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.165987 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="mysql-bootstrap" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166002 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166009 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166023 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166030 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166042 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166050 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166063 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="extract-utilities" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166071 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="extract-utilities" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166085 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" containerName="nova-cell0-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166093 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" containerName="nova-cell0-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166105 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166113 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166125 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166132 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166144 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166151 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166164 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" containerName="memcached" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166172 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" containerName="memcached" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166182 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166190 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166199 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="cinder-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166208 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="cinder-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166221 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5e54c4-ad98-4f26-9681-189997f3bd64" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166228 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5e54c4-ad98-4f26-9681-189997f3bd64" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166240 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="mysql-bootstrap" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166247 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="mysql-bootstrap" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166254 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="sg-core" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166261 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="sg-core" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166270 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-central-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166280 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-central-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166295 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166302 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166314 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-notification-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166322 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-notification-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166337 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166344 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166356 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="setup-container" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166364 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="setup-container" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166372 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166379 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166390 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166399 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166413 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166421 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166438 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="init" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166445 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="init" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166460 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166468 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166482 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2851153-f304-4041-9de2-045c0d92526d" containerName="kube-state-metrics" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166489 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2851153-f304-4041-9de2-045c0d92526d" containerName="kube-state-metrics" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166498 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166505 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166520 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="ovsdbserver-nb" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166528 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="ovsdbserver-nb" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166541 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166548 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166561 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166569 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166578 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166585 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166598 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166605 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166613 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="setup-container" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166620 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="setup-container" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166634 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3bb280-8572-448d-a3af-4c33993fc232" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166641 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3bb280-8572-448d-a3af-4c33993fc232" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166653 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server-init" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166660 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server-init" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166672 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="rsync" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166679 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="rsync" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166709 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166733 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166747 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873cb35-815f-44ae-b38d-bb7ceb806b9d" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166755 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873cb35-815f-44ae-b38d-bb7ceb806b9d" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166768 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166775 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166785 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166794 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166803 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166810 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166817 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166825 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166834 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166842 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166854 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166861 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166869 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166877 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166893 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerName="nova-scheduler-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166900 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerName="nova-scheduler-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166914 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="swift-recon-cron" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166921 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="swift-recon-cron" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166935 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166942 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166956 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166963 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166972 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="943a976f-044a-4b42-b08c-baeae7b85f42" containerName="keystone-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166980 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="943a976f-044a-4b42-b08c-baeae7b85f42" containerName="keystone-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.166990 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.166997 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167008 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167016 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167026 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167034 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167046 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="extract-content" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167054 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="extract-content" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167067 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167074 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167085 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="probe" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167092 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="probe" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167104 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167111 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167119 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167128 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167137 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce4369c-e009-4ae1-bc07-65dff0005807" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167145 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce4369c-e009-4ae1-bc07-65dff0005807" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167156 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="registry-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167163 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="registry-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167172 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167179 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167190 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167197 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167207 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="ovsdbserver-sb" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167215 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="ovsdbserver-sb" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167228 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167236 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167247 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167255 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167268 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167276 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167287 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-expirer" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167294 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-expirer" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167302 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167310 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167320 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167327 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-server" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167341 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167349 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167357 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-reaper" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167365 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-reaper" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167376 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f72fdf5-3656-49e2-80a8-37d0d65ec212" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167384 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f72fdf5-3656-49e2-80a8-37d0d65ec212" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167393 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167401 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167411 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167418 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167428 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595b1839-8525-4b91-ab30-fd800710cf31" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167438 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="595b1839-8525-4b91-ab30-fd800710cf31" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: E1001 07:30:00.167453 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfefe16-d728-4570-9ef6-cc717162f590" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167462 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfefe16-d728-4570-9ef6-cc717162f590" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167610 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-notification-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167626 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167637 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovs-vswitchd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167649 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="rsync" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167657 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167666 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddfefe16-d728-4570-9ef6-cc717162f590" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167675 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5e54c4-ad98-4f26-9681-189997f3bd64" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167686 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce4369c-e009-4ae1-bc07-65dff0005807" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167721 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="595b1839-8525-4b91-ab30-fd800710cf31" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167734 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167746 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167760 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167771 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="28035a28-28d4-46b3-b269-97a523cf731e" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167780 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6063ad95-cbaf-4798-899f-195d4428f246" containerName="nova-metadata-metadata" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167791 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="ovsdbserver-sb" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167804 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="06af3d93-c582-4d5c-97d0-6a0250b1eb10" containerName="barbican-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167815 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e5feff-74ce-434a-8606-85c51844deb9" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167829 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167841 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f3dfeb8-859c-46d2-a7a2-17ae34ebc9da" containerName="barbican-worker" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167853 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167865 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167873 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167884 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="swift-recon-cron" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167893 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="ovsdbserver-nb" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167905 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167916 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="295f86b5-4cd2-4ceb-bf30-b3397a6f5f62" containerName="ovn-controller" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167926 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f871dbb-1e64-411b-acfa-39ed6b5b66ed" containerName="barbican-keystone-listener" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167936 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb9ef6d-db8f-4747-b93c-fa927429a785" containerName="ovsdb-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167955 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167966 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f2490fe-759d-4263-ae63-0f69c4f78c7d" containerName="memcached" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167982 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f72fdf5-3656-49e2-80a8-37d0d65ec212" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.167998 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-expirer" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168011 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="144c538f-4172-4043-b716-62f91b64f8fc" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168022 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46a0ffe-6c41-4d34-bbe9-8db505bbe46b" containerName="nova-cell1-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168038 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168049 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee312865-a3af-43d5-9b7d-13d62695af31" containerName="openstack-network-exporter" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168064 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168075 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c09c611-b859-4650-b4b0-96738b65fc45" containerName="nova-cell0-conductor-conductor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168092 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a848f4-7b85-4463-bdae-f763e6f9e7e5" containerName="dnsmasq-dns" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168103 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="probe" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168120 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2339ab9b-c72a-4b13-81f8-61b6a9d6e4a8" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168134 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-updater" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168143 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="08552fbf-b050-46ec-b8bf-83cde04f87e6" containerName="nova-api-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168152 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dcde73a-d928-45af-914d-3b1f5f7f899b" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168162 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168172 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="ceilometer-central-agent" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168182 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2851153-f304-4041-9de2-045c0d92526d" containerName="kube-state-metrics" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168195 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168209 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="account-reaper" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168221 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168232 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168245 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39fc78-2eb7-4a57-8bbc-11733cbb366a" containerName="proxy-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168255 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168267 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e3523b-a91a-4d3d-85c7-b41e4006c0e2" containerName="neutron-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168278 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-log" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168287 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="23126d10-0ad7-45e9-94e3-3e34116c20d8" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168296 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e65e7c-ced2-4e38-9841-3f383471b275" containerName="glance-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168306 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873cb35-815f-44ae-b38d-bb7ceb806b9d" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168321 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="container-auditor" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168328 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-replicator" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168336 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="proxy-httpd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168345 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f15a72f-f4ec-4e07-ae6d-7eef5700f543" containerName="nova-scheduler-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168355 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27ce0b1-38ff-4a33-be51-5ba651e8e075" containerName="object-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168365 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ff50da-fcbe-4fe9-9a35-073f2abdc3e7" containerName="cinder-scheduler" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168375 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="943a976f-044a-4b42-b08c-baeae7b85f42" containerName="keystone-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168382 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d64002d-6a47-4928-9495-91a58f4b2a5c" containerName="cinder-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168392 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="17575b0a-a27e-4cc5-a84a-0ed9adf0ed65" containerName="sg-core" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168401 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e273fd30-371c-494e-99ce-6460b2a1dd00" containerName="registry-server" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168411 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6212cb23-af83-4bcf-8328-91b14a62844b" containerName="placement-api" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168422 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3bb280-8572-448d-a3af-4c33993fc232" containerName="mariadb-account-delete" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168431 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="44374bf9-c936-4c06-ac74-e80244c09039" containerName="galera" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168442 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4140f175-0421-48d7-84e1-11fd6cdcea52" containerName="ovn-northd" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.168452 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd1fb4b-b175-4ed4-8e14-05c2a5fe3087" containerName="rabbitmq" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.169005 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.171799 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.172309 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.194639 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj"] Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.263633 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.263840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bt6\" (UniqueName: \"kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.263906 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.365849 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bt6\" (UniqueName: \"kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.365993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.366086 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.368094 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.384360 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bt6\" (UniqueName: \"kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.388880 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume\") pod \"collect-profiles-29321730-ldphj\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.499738 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:00 crc kubenswrapper[4837]: I1001 07:30:00.942930 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj"] Oct 01 07:30:01 crc kubenswrapper[4837]: I1001 07:30:01.698145 4837 generic.go:334] "Generic (PLEG): container finished" podID="add6774c-ba3a-4da8-bc3b-ee0db048329b" containerID="f8473a10b72c7d997561e536f90d2dacf3e5c5011a16e107138e5123ae45e42d" exitCode=0 Oct 01 07:30:01 crc kubenswrapper[4837]: I1001 07:30:01.698253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" event={"ID":"add6774c-ba3a-4da8-bc3b-ee0db048329b","Type":"ContainerDied","Data":"f8473a10b72c7d997561e536f90d2dacf3e5c5011a16e107138e5123ae45e42d"} Oct 01 07:30:01 crc kubenswrapper[4837]: I1001 07:30:01.698458 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" event={"ID":"add6774c-ba3a-4da8-bc3b-ee0db048329b","Type":"ContainerStarted","Data":"6690f399583870715a4775bf15b4b604c51b509e21bb57d259e55c7799e4af35"} Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.056045 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.207540 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume\") pod \"add6774c-ba3a-4da8-bc3b-ee0db048329b\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.207631 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bt6\" (UniqueName: \"kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6\") pod \"add6774c-ba3a-4da8-bc3b-ee0db048329b\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.207675 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume\") pod \"add6774c-ba3a-4da8-bc3b-ee0db048329b\" (UID: \"add6774c-ba3a-4da8-bc3b-ee0db048329b\") " Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.208463 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume" (OuterVolumeSpecName: "config-volume") pod "add6774c-ba3a-4da8-bc3b-ee0db048329b" (UID: "add6774c-ba3a-4da8-bc3b-ee0db048329b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.216316 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6" (OuterVolumeSpecName: "kube-api-access-97bt6") pod "add6774c-ba3a-4da8-bc3b-ee0db048329b" (UID: "add6774c-ba3a-4da8-bc3b-ee0db048329b"). InnerVolumeSpecName "kube-api-access-97bt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.221792 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "add6774c-ba3a-4da8-bc3b-ee0db048329b" (UID: "add6774c-ba3a-4da8-bc3b-ee0db048329b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.309548 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/add6774c-ba3a-4da8-bc3b-ee0db048329b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.309599 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/add6774c-ba3a-4da8-bc3b-ee0db048329b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.309620 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bt6\" (UniqueName: \"kubernetes.io/projected/add6774c-ba3a-4da8-bc3b-ee0db048329b-kube-api-access-97bt6\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.718729 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" event={"ID":"add6774c-ba3a-4da8-bc3b-ee0db048329b","Type":"ContainerDied","Data":"6690f399583870715a4775bf15b4b604c51b509e21bb57d259e55c7799e4af35"} Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.719072 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6690f399583870715a4775bf15b4b604c51b509e21bb57d259e55c7799e4af35" Oct 01 07:30:03 crc kubenswrapper[4837]: I1001 07:30:03.719136 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.321511 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:14 crc kubenswrapper[4837]: E1001 07:30:14.322567 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="add6774c-ba3a-4da8-bc3b-ee0db048329b" containerName="collect-profiles" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.322590 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="add6774c-ba3a-4da8-bc3b-ee0db048329b" containerName="collect-profiles" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.322921 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="add6774c-ba3a-4da8-bc3b-ee0db048329b" containerName="collect-profiles" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.324756 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.337520 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.485639 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.485781 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7fm2\" (UniqueName: \"kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.485827 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.587256 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7fm2\" (UniqueName: \"kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.587353 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.587536 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.588460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.588560 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.646083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7fm2\" (UniqueName: \"kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2\") pod \"redhat-marketplace-hvh5r\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:14 crc kubenswrapper[4837]: I1001 07:30:14.660208 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:15 crc kubenswrapper[4837]: I1001 07:30:15.130543 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:15 crc kubenswrapper[4837]: W1001 07:30:15.143795 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8efbfb67_3b31_4bc9_8136_258d6a3d50bd.slice/crio-174a8dafa401570a1b2f3a09ec22fe996966ab31b0677bb8e6b258645e2ba041 WatchSource:0}: Error finding container 174a8dafa401570a1b2f3a09ec22fe996966ab31b0677bb8e6b258645e2ba041: Status 404 returned error can't find the container with id 174a8dafa401570a1b2f3a09ec22fe996966ab31b0677bb8e6b258645e2ba041 Oct 01 07:30:15 crc kubenswrapper[4837]: I1001 07:30:15.849015 4837 generic.go:334] "Generic (PLEG): container finished" podID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerID="445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf" exitCode=0 Oct 01 07:30:15 crc kubenswrapper[4837]: I1001 07:30:15.849082 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerDied","Data":"445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf"} Oct 01 07:30:15 crc kubenswrapper[4837]: I1001 07:30:15.849484 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerStarted","Data":"174a8dafa401570a1b2f3a09ec22fe996966ab31b0677bb8e6b258645e2ba041"} Oct 01 07:30:17 crc kubenswrapper[4837]: I1001 07:30:17.887583 4837 generic.go:334] "Generic (PLEG): container finished" podID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerID="8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91" exitCode=0 Oct 01 07:30:17 crc kubenswrapper[4837]: I1001 07:30:17.887706 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerDied","Data":"8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91"} Oct 01 07:30:18 crc kubenswrapper[4837]: I1001 07:30:18.910027 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerStarted","Data":"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e"} Oct 01 07:30:18 crc kubenswrapper[4837]: I1001 07:30:18.940604 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hvh5r" podStartSLOduration=2.434260146 podStartE2EDuration="4.940587171s" podCreationTimestamp="2025-10-01 07:30:14 +0000 UTC" firstStartedPulling="2025-10-01 07:30:15.851669768 +0000 UTC m=+1472.693277253" lastFinishedPulling="2025-10-01 07:30:18.357996823 +0000 UTC m=+1475.199604278" observedRunningTime="2025-10-01 07:30:18.939917635 +0000 UTC m=+1475.781525100" watchObservedRunningTime="2025-10-01 07:30:18.940587171 +0000 UTC m=+1475.782194636" Oct 01 07:30:23 crc kubenswrapper[4837]: I1001 07:30:23.079808 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:30:23 crc kubenswrapper[4837]: I1001 07:30:23.080155 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:30:24 crc kubenswrapper[4837]: I1001 07:30:24.660436 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:24 crc kubenswrapper[4837]: I1001 07:30:24.660846 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:24 crc kubenswrapper[4837]: I1001 07:30:24.740626 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:25 crc kubenswrapper[4837]: I1001 07:30:25.055777 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:25 crc kubenswrapper[4837]: I1001 07:30:25.125082 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.003246 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hvh5r" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="registry-server" containerID="cri-o://a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e" gracePeriod=2 Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.497304 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.625860 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7fm2\" (UniqueName: \"kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2\") pod \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.626024 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities\") pod \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.626087 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content\") pod \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\" (UID: \"8efbfb67-3b31-4bc9-8136-258d6a3d50bd\") " Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.627079 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities" (OuterVolumeSpecName: "utilities") pod "8efbfb67-3b31-4bc9-8136-258d6a3d50bd" (UID: "8efbfb67-3b31-4bc9-8136-258d6a3d50bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.631244 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2" (OuterVolumeSpecName: "kube-api-access-w7fm2") pod "8efbfb67-3b31-4bc9-8136-258d6a3d50bd" (UID: "8efbfb67-3b31-4bc9-8136-258d6a3d50bd"). InnerVolumeSpecName "kube-api-access-w7fm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.652324 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8efbfb67-3b31-4bc9-8136-258d6a3d50bd" (UID: "8efbfb67-3b31-4bc9-8136-258d6a3d50bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.727503 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7fm2\" (UniqueName: \"kubernetes.io/projected/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-kube-api-access-w7fm2\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.727558 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:27 crc kubenswrapper[4837]: I1001 07:30:27.727573 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8efbfb67-3b31-4bc9-8136-258d6a3d50bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.016012 4837 generic.go:334] "Generic (PLEG): container finished" podID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerID="a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e" exitCode=0 Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.016036 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hvh5r" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.016064 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerDied","Data":"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e"} Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.016126 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hvh5r" event={"ID":"8efbfb67-3b31-4bc9-8136-258d6a3d50bd","Type":"ContainerDied","Data":"174a8dafa401570a1b2f3a09ec22fe996966ab31b0677bb8e6b258645e2ba041"} Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.016155 4837 scope.go:117] "RemoveContainer" containerID="a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.065980 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.076722 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hvh5r"] Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.087043 4837 scope.go:117] "RemoveContainer" containerID="8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.114645 4837 scope.go:117] "RemoveContainer" containerID="445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.158102 4837 scope.go:117] "RemoveContainer" containerID="a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e" Oct 01 07:30:28 crc kubenswrapper[4837]: E1001 07:30:28.159031 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e\": container with ID starting with a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e not found: ID does not exist" containerID="a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.159083 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e"} err="failed to get container status \"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e\": rpc error: code = NotFound desc = could not find container \"a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e\": container with ID starting with a578c3920c457448bfbb87eda2a0ca514837f7b851d43974546e2cfc05558c7e not found: ID does not exist" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.159125 4837 scope.go:117] "RemoveContainer" containerID="8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91" Oct 01 07:30:28 crc kubenswrapper[4837]: E1001 07:30:28.159641 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91\": container with ID starting with 8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91 not found: ID does not exist" containerID="8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.159798 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91"} err="failed to get container status \"8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91\": rpc error: code = NotFound desc = could not find container \"8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91\": container with ID starting with 8f4c81c4f2a5f63273068967b550e18aa8383f06207ee9d4cbe0ff7d5153ff91 not found: ID does not exist" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.159866 4837 scope.go:117] "RemoveContainer" containerID="445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf" Oct 01 07:30:28 crc kubenswrapper[4837]: E1001 07:30:28.160302 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf\": container with ID starting with 445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf not found: ID does not exist" containerID="445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf" Oct 01 07:30:28 crc kubenswrapper[4837]: I1001 07:30:28.160370 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf"} err="failed to get container status \"445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf\": rpc error: code = NotFound desc = could not find container \"445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf\": container with ID starting with 445717ad5d4840b832a248fbea587b3a74ea1136b926afe7648de791afcbb6bf not found: ID does not exist" Oct 01 07:30:29 crc kubenswrapper[4837]: I1001 07:30:29.830337 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" path="/var/lib/kubelet/pods/8efbfb67-3b31-4bc9-8136-258d6a3d50bd/volumes" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.080358 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.081122 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.081196 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.082238 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.082361 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" gracePeriod=600 Oct 01 07:30:53 crc kubenswrapper[4837]: E1001 07:30:53.209819 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.290185 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" exitCode=0 Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.290249 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d"} Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.290303 4837 scope.go:117] "RemoveContainer" containerID="49facd65b68b6336661ccd0f232f60a32185ad1fc4ef38671a44e2e94f580e92" Oct 01 07:30:53 crc kubenswrapper[4837]: I1001 07:30:53.290996 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:30:53 crc kubenswrapper[4837]: E1001 07:30:53.291379 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.485232 4837 scope.go:117] "RemoveContainer" containerID="26e6a34c5516afcbbf5945753d417ac28fae4221a6a145124aa496a1fd01b8b4" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.546002 4837 scope.go:117] "RemoveContainer" containerID="3604553f9371c79bc1695100ccdedabfe3a3e93d224aaa7de52af1ec682e6662" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.596136 4837 scope.go:117] "RemoveContainer" containerID="af518aa79be249c5dd079020e8fd4ce5155844a7807fbfb5824990405200f7d4" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.639296 4837 scope.go:117] "RemoveContainer" containerID="a98f366483806d112bcd2392d65b5f235d68f2b5b319dc8a8571c949a57bb438" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.695034 4837 scope.go:117] "RemoveContainer" containerID="3aa66d386d064c0c3f0f19d91744583e91dad1847715857b333c84b42876325a" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.717570 4837 scope.go:117] "RemoveContainer" containerID="c0eefaff1b420ae4a1c96b3aeb67603ed36b3c9ecd70481297dfbf7c87570460" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.735669 4837 scope.go:117] "RemoveContainer" containerID="81bc85eef741d4dce8bac1d030483f50daf2acf2f5c76f67467319810a845025" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.758876 4837 scope.go:117] "RemoveContainer" containerID="6d9db3522d4f0e2faad737251daa28edf3802cc6ddbfe74acf2694f2acb1b8ac" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.781084 4837 scope.go:117] "RemoveContainer" containerID="a622535f33e24a5a3d6940ec66cb10d593f27ec6d5ccecd06dc67b1ffadadcf0" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.797733 4837 scope.go:117] "RemoveContainer" containerID="e975bd72ab746ac62c6e87b203ec843569239bcdf63e46adc77064a72d2c8af5" Oct 01 07:31:00 crc kubenswrapper[4837]: I1001 07:31:00.814187 4837 scope.go:117] "RemoveContainer" containerID="81870eeb531bc2fdf0c9e9a1cb9ae49dea463978c323ba7711a80e39cc7b2d2e" Oct 01 07:31:05 crc kubenswrapper[4837]: I1001 07:31:05.816200 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:31:05 crc kubenswrapper[4837]: E1001 07:31:05.816874 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:31:18 crc kubenswrapper[4837]: I1001 07:31:18.816167 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:31:18 crc kubenswrapper[4837]: E1001 07:31:18.817456 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:31:31 crc kubenswrapper[4837]: I1001 07:31:31.816873 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:31:31 crc kubenswrapper[4837]: E1001 07:31:31.818361 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:31:42 crc kubenswrapper[4837]: I1001 07:31:42.816193 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:31:42 crc kubenswrapper[4837]: E1001 07:31:42.818610 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:31:57 crc kubenswrapper[4837]: I1001 07:31:57.816144 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:31:57 crc kubenswrapper[4837]: E1001 07:31:57.817105 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.046323 4837 scope.go:117] "RemoveContainer" containerID="3a08d66e194419e08949044103a9a74620bbcd529385fbab247b8a2c78c247bd" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.102622 4837 scope.go:117] "RemoveContainer" containerID="dd9dba269473c23311be1d264a0a8a5a066fb07a7e2e6a62e54478bad4794813" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.131035 4837 scope.go:117] "RemoveContainer" containerID="31c1c4531205c7aa377e609c7ac5c5f31a4a0e827612da211359083203b511a5" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.152538 4837 scope.go:117] "RemoveContainer" containerID="f79c5da489b3e4e2a44ecd79261d14e1290a38a0e382e01f266b6136c16b1250" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.195333 4837 scope.go:117] "RemoveContainer" containerID="93da5dd64723a812493ca4183a35399159dcfacc340b861719f66b99aa3185a3" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.220297 4837 scope.go:117] "RemoveContainer" containerID="1ba5135a56f71ec6e35ec8aca0609be3d9cd38f58451416cc4bbea13198a11fe" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.259678 4837 scope.go:117] "RemoveContainer" containerID="19c640a121f67ee824cde93f790747e72586eaf1ece8a807acdd3a97e2cbe3ac" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.280608 4837 scope.go:117] "RemoveContainer" containerID="f0d02bee9b43b1aa4c0d88fee58fc197ff6560dc60daaca70139c6751c5ce0a9" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.310837 4837 scope.go:117] "RemoveContainer" containerID="986d92bb1fcb5055d52a0dfa057171bd964fd3fe435bf9c633eba6301866d672" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.330324 4837 scope.go:117] "RemoveContainer" containerID="c3b73bdd602f1754beef682cc6ae2997909ff57156cad987858a8ff3e196d7a8" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.360268 4837 scope.go:117] "RemoveContainer" containerID="0fc6fdf9472d712ace5bbb580bcdc39380a6fc356087908b1642817b1194f4c7" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.398539 4837 scope.go:117] "RemoveContainer" containerID="5d7fc883157d22705620cceb566a7750f43040d9aac88d9b6d381c18b7520ea5" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.433834 4837 scope.go:117] "RemoveContainer" containerID="2f1b15f1d32699a3f4010b989a02711f339c01163b9b29fe3a66969c2d8cdcff" Oct 01 07:32:01 crc kubenswrapper[4837]: I1001 07:32:01.457714 4837 scope.go:117] "RemoveContainer" containerID="a4bd3be67a6248edc35435d40996f9d6bc93d305b11058483bb4679159edd1be" Oct 01 07:32:10 crc kubenswrapper[4837]: I1001 07:32:10.815443 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:32:10 crc kubenswrapper[4837]: E1001 07:32:10.816491 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:32:25 crc kubenswrapper[4837]: I1001 07:32:25.816144 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:32:25 crc kubenswrapper[4837]: E1001 07:32:25.817257 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:32:40 crc kubenswrapper[4837]: I1001 07:32:40.816651 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:32:40 crc kubenswrapper[4837]: E1001 07:32:40.817947 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:32:53 crc kubenswrapper[4837]: I1001 07:32:53.824534 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:32:53 crc kubenswrapper[4837]: E1001 07:32:53.825686 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.710891 4837 scope.go:117] "RemoveContainer" containerID="f2f44d90065f3b09e2e0423a1cb4fbaf07ef26a225a477036e37641545512570" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.754116 4837 scope.go:117] "RemoveContainer" containerID="9dfecee7787a639e916f3c2e6d2900c5192cf19d000491e365456ae7ad50656d" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.816089 4837 scope.go:117] "RemoveContainer" containerID="e8db386829c99e752736e8a3a56fdc0ddce5b550d7fd78fdcfb1d074bd8ca628" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.840335 4837 scope.go:117] "RemoveContainer" containerID="e7843f1bad6c38139b08c3822653e6cdb5f57f90c6ed5dd10603af9b3c5fb0aa" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.860334 4837 scope.go:117] "RemoveContainer" containerID="3552d4b1a01b53a4ee56dd82ae9cca651bd961e33681b9c72d8f9260300ac389" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.891974 4837 scope.go:117] "RemoveContainer" containerID="cdd54b9fd4aba5b9af8e7129caeebcabaf6de2161132e3fe487197e5759bae78" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.921559 4837 scope.go:117] "RemoveContainer" containerID="12ec459428014c7466214d2c99d3446b10ae3bcac80009aa60ef4f936e7fd003" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.951035 4837 scope.go:117] "RemoveContainer" containerID="8453b03acddb5b9c4dbf1558763948d0e396c1231e9e3ea725a0e06c4ac8956d" Oct 01 07:33:01 crc kubenswrapper[4837]: I1001 07:33:01.981181 4837 scope.go:117] "RemoveContainer" containerID="f94c2451dd653c31792c302868b2003480f76be40f9c075f3e63a6788149c348" Oct 01 07:33:02 crc kubenswrapper[4837]: I1001 07:33:02.047483 4837 scope.go:117] "RemoveContainer" containerID="d87ec853871ad0d44a4f50125b836c1f87633551c852c805248ea18a8a04fa83" Oct 01 07:33:02 crc kubenswrapper[4837]: I1001 07:33:02.077056 4837 scope.go:117] "RemoveContainer" containerID="9d0c63a56b505c323247613cb05d39bfe8227df3c32c1ff6ebcabfa1e90233e6" Oct 01 07:33:02 crc kubenswrapper[4837]: I1001 07:33:02.109554 4837 scope.go:117] "RemoveContainer" containerID="b7792f2f321c19d1eb1568ae96eddc87b77661b3d811ceac5cfe3f464d5b9726" Oct 01 07:33:02 crc kubenswrapper[4837]: I1001 07:33:02.137783 4837 scope.go:117] "RemoveContainer" containerID="d863c9fe0ae9e19204566c6cb4300be788afaf2105039c052930acb1e3a11e23" Oct 01 07:33:02 crc kubenswrapper[4837]: I1001 07:33:02.172311 4837 scope.go:117] "RemoveContainer" containerID="c5de51b5f6c2a76c7e0d181f7cdfb7392fdc8323c0f38c8d9eaf84c7dfe54255" Oct 01 07:33:05 crc kubenswrapper[4837]: I1001 07:33:05.816745 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:33:05 crc kubenswrapper[4837]: E1001 07:33:05.817432 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:33:18 crc kubenswrapper[4837]: I1001 07:33:18.815907 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:33:18 crc kubenswrapper[4837]: E1001 07:33:18.817350 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:33:31 crc kubenswrapper[4837]: I1001 07:33:31.816202 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:33:31 crc kubenswrapper[4837]: E1001 07:33:31.817205 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:33:46 crc kubenswrapper[4837]: I1001 07:33:46.816344 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:33:46 crc kubenswrapper[4837]: E1001 07:33:46.817619 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:00 crc kubenswrapper[4837]: I1001 07:34:00.816566 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:34:00 crc kubenswrapper[4837]: E1001 07:34:00.817588 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:02 crc kubenswrapper[4837]: I1001 07:34:02.453784 4837 scope.go:117] "RemoveContainer" containerID="7e1bae08709298165b16e91922f8d331270e19bce0e857867200a431bc764409" Oct 01 07:34:02 crc kubenswrapper[4837]: I1001 07:34:02.486342 4837 scope.go:117] "RemoveContainer" containerID="8b0403847fa1e913e31eb2c874d46883eb7f9de350b911b8a7adf2ee314d096a" Oct 01 07:34:02 crc kubenswrapper[4837]: I1001 07:34:02.547868 4837 scope.go:117] "RemoveContainer" containerID="034af4cc06f98a5a1c1bf69b7c633186fde48f42cccb9b972d748a605cd9eec8" Oct 01 07:34:02 crc kubenswrapper[4837]: I1001 07:34:02.578995 4837 scope.go:117] "RemoveContainer" containerID="af16c63816c1328370401333ff150789d7df93335939e3bb0c6ec880848f98b8" Oct 01 07:34:02 crc kubenswrapper[4837]: I1001 07:34:02.636940 4837 scope.go:117] "RemoveContainer" containerID="8b0122a9f0b165396dc68834bcd670b9d796dd63c3cd41cd5cc6e187caf12ecf" Oct 01 07:34:14 crc kubenswrapper[4837]: I1001 07:34:14.815584 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:34:14 crc kubenswrapper[4837]: E1001 07:34:14.816767 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:25 crc kubenswrapper[4837]: I1001 07:34:25.816308 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:34:25 crc kubenswrapper[4837]: E1001 07:34:25.816987 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:38 crc kubenswrapper[4837]: I1001 07:34:38.816408 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:34:38 crc kubenswrapper[4837]: E1001 07:34:38.817425 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.918017 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:34:50 crc kubenswrapper[4837]: E1001 07:34:50.919137 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="registry-server" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.919159 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="registry-server" Oct 01 07:34:50 crc kubenswrapper[4837]: E1001 07:34:50.919185 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="extract-utilities" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.919198 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="extract-utilities" Oct 01 07:34:50 crc kubenswrapper[4837]: E1001 07:34:50.919218 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="extract-content" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.919232 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="extract-content" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.919536 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8efbfb67-3b31-4bc9-8136-258d6a3d50bd" containerName="registry-server" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.921828 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:50 crc kubenswrapper[4837]: I1001 07:34:50.938438 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.073014 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkm5h\" (UniqueName: \"kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.073427 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.073595 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.175093 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.175509 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.175857 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkm5h\" (UniqueName: \"kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.176048 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.176425 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.214600 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkm5h\" (UniqueName: \"kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h\") pod \"certified-operators-n9gsx\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.284315 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.555751 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.773991 4837 generic.go:334] "Generic (PLEG): container finished" podID="faeb9800-03fb-4c54-9eb3-880238391caf" containerID="b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff" exitCode=0 Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.774042 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerDied","Data":"b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff"} Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.774280 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerStarted","Data":"9e98f8f1aab9ceb19c66c68763867a6142292c07afded956435c1a033d9bd3b0"} Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.775570 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:34:51 crc kubenswrapper[4837]: I1001 07:34:51.816559 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:34:51 crc kubenswrapper[4837]: E1001 07:34:51.817005 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:34:52 crc kubenswrapper[4837]: I1001 07:34:52.787242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerStarted","Data":"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0"} Oct 01 07:34:53 crc kubenswrapper[4837]: I1001 07:34:53.801095 4837 generic.go:334] "Generic (PLEG): container finished" podID="faeb9800-03fb-4c54-9eb3-880238391caf" containerID="7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0" exitCode=0 Oct 01 07:34:53 crc kubenswrapper[4837]: I1001 07:34:53.801154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerDied","Data":"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0"} Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.109862 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ln2gs"] Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.111769 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.126297 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ln2gs"] Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.238679 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl2qg\" (UniqueName: \"kubernetes.io/projected/82ec76e8-dee7-47ad-be82-6eca642819b3-kube-api-access-pl2qg\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.238798 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-utilities\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.238886 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-catalog-content\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.339816 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl2qg\" (UniqueName: \"kubernetes.io/projected/82ec76e8-dee7-47ad-be82-6eca642819b3-kube-api-access-pl2qg\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.339913 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-utilities\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.339963 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-catalog-content\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.340528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-utilities\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.340571 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82ec76e8-dee7-47ad-be82-6eca642819b3-catalog-content\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.363914 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl2qg\" (UniqueName: \"kubernetes.io/projected/82ec76e8-dee7-47ad-be82-6eca642819b3-kube-api-access-pl2qg\") pod \"community-operators-ln2gs\" (UID: \"82ec76e8-dee7-47ad-be82-6eca642819b3\") " pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.443074 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.714740 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ln2gs"] Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.808947 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerStarted","Data":"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127"} Oct 01 07:34:54 crc kubenswrapper[4837]: I1001 07:34:54.811829 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ln2gs" event={"ID":"82ec76e8-dee7-47ad-be82-6eca642819b3","Type":"ContainerStarted","Data":"3e164da5cf45970f4c7932a286a3464310913f509058c94b1af3b790c21a9f93"} Oct 01 07:34:55 crc kubenswrapper[4837]: I1001 07:34:55.824539 4837 generic.go:334] "Generic (PLEG): container finished" podID="82ec76e8-dee7-47ad-be82-6eca642819b3" containerID="25f428beae15c257eef12f869fda5ec679797d2285f264166589e39408a33ac3" exitCode=0 Oct 01 07:34:55 crc kubenswrapper[4837]: I1001 07:34:55.832475 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ln2gs" event={"ID":"82ec76e8-dee7-47ad-be82-6eca642819b3","Type":"ContainerDied","Data":"25f428beae15c257eef12f869fda5ec679797d2285f264166589e39408a33ac3"} Oct 01 07:34:55 crc kubenswrapper[4837]: I1001 07:34:55.852170 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n9gsx" podStartSLOduration=3.4147237329999998 podStartE2EDuration="5.852131905s" podCreationTimestamp="2025-10-01 07:34:50 +0000 UTC" firstStartedPulling="2025-10-01 07:34:51.775355779 +0000 UTC m=+1748.616963234" lastFinishedPulling="2025-10-01 07:34:54.212763951 +0000 UTC m=+1751.054371406" observedRunningTime="2025-10-01 07:34:54.826084772 +0000 UTC m=+1751.667692227" watchObservedRunningTime="2025-10-01 07:34:55.852131905 +0000 UTC m=+1752.693739370" Oct 01 07:34:59 crc kubenswrapper[4837]: I1001 07:34:59.876659 4837 generic.go:334] "Generic (PLEG): container finished" podID="82ec76e8-dee7-47ad-be82-6eca642819b3" containerID="26c5bde3a9fdf969f1c3c685d8183c5511617b1d2c18d380ae81701cb684c6b1" exitCode=0 Oct 01 07:34:59 crc kubenswrapper[4837]: I1001 07:34:59.876755 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ln2gs" event={"ID":"82ec76e8-dee7-47ad-be82-6eca642819b3","Type":"ContainerDied","Data":"26c5bde3a9fdf969f1c3c685d8183c5511617b1d2c18d380ae81701cb684c6b1"} Oct 01 07:35:00 crc kubenswrapper[4837]: I1001 07:35:00.885916 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ln2gs" event={"ID":"82ec76e8-dee7-47ad-be82-6eca642819b3","Type":"ContainerStarted","Data":"29e32df02fbca66be6bc4a5acd1928abbe67fc4a83c214845c55387fe52bd015"} Oct 01 07:35:01 crc kubenswrapper[4837]: I1001 07:35:01.285519 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:01 crc kubenswrapper[4837]: I1001 07:35:01.286191 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:01 crc kubenswrapper[4837]: I1001 07:35:01.347597 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:01 crc kubenswrapper[4837]: I1001 07:35:01.384429 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ln2gs" podStartSLOduration=2.830128996 podStartE2EDuration="7.384383239s" podCreationTimestamp="2025-10-01 07:34:54 +0000 UTC" firstStartedPulling="2025-10-01 07:34:55.827616692 +0000 UTC m=+1752.669224177" lastFinishedPulling="2025-10-01 07:35:00.381870925 +0000 UTC m=+1757.223478420" observedRunningTime="2025-10-01 07:35:00.905242622 +0000 UTC m=+1757.746850077" watchObservedRunningTime="2025-10-01 07:35:01.384383239 +0000 UTC m=+1758.225990704" Oct 01 07:35:01 crc kubenswrapper[4837]: I1001 07:35:01.969194 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.500653 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.793388 4837 scope.go:117] "RemoveContainer" containerID="ed6ba9d9762a04643429cdf39156487cb7939eb11057b92327a090e24655643f" Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.851516 4837 scope.go:117] "RemoveContainer" containerID="e16cb98018d7c03dd0dd1ecd72811712a0e0dd054524e8cbf1d1ad4808a82d5a" Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.876865 4837 scope.go:117] "RemoveContainer" containerID="60a035800ab40246518e36001a9d37d48e3accfcf08a7a2ff8966dd3e863a8e4" Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.906622 4837 scope.go:117] "RemoveContainer" containerID="82f7d563648b7707832c5c7466bc63bb736e1c4e59613467ee367330271cad6d" Oct 01 07:35:02 crc kubenswrapper[4837]: I1001 07:35:02.942372 4837 scope.go:117] "RemoveContainer" containerID="5ed2be1636cd3ecb4b1b28b8de77bab8aa138fe6f32f2d417e861cbe3e2e5a4a" Oct 01 07:35:03 crc kubenswrapper[4837]: I1001 07:35:03.921093 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n9gsx" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="registry-server" containerID="cri-o://fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127" gracePeriod=2 Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.406141 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.454791 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.454840 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.503347 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.595666 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkm5h\" (UniqueName: \"kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h\") pod \"faeb9800-03fb-4c54-9eb3-880238391caf\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.595767 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content\") pod \"faeb9800-03fb-4c54-9eb3-880238391caf\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.595921 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities\") pod \"faeb9800-03fb-4c54-9eb3-880238391caf\" (UID: \"faeb9800-03fb-4c54-9eb3-880238391caf\") " Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.598044 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities" (OuterVolumeSpecName: "utilities") pod "faeb9800-03fb-4c54-9eb3-880238391caf" (UID: "faeb9800-03fb-4c54-9eb3-880238391caf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.604656 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h" (OuterVolumeSpecName: "kube-api-access-dkm5h") pod "faeb9800-03fb-4c54-9eb3-880238391caf" (UID: "faeb9800-03fb-4c54-9eb3-880238391caf"). InnerVolumeSpecName "kube-api-access-dkm5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.674591 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "faeb9800-03fb-4c54-9eb3-880238391caf" (UID: "faeb9800-03fb-4c54-9eb3-880238391caf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.698275 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkm5h\" (UniqueName: \"kubernetes.io/projected/faeb9800-03fb-4c54-9eb3-880238391caf-kube-api-access-dkm5h\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.698339 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.698363 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faeb9800-03fb-4c54-9eb3-880238391caf-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.815852 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:35:04 crc kubenswrapper[4837]: E1001 07:35:04.816537 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.933985 4837 generic.go:334] "Generic (PLEG): container finished" podID="faeb9800-03fb-4c54-9eb3-880238391caf" containerID="fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127" exitCode=0 Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.934135 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9gsx" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.934113 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerDied","Data":"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127"} Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.934352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9gsx" event={"ID":"faeb9800-03fb-4c54-9eb3-880238391caf","Type":"ContainerDied","Data":"9e98f8f1aab9ceb19c66c68763867a6142292c07afded956435c1a033d9bd3b0"} Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.934393 4837 scope.go:117] "RemoveContainer" containerID="fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.961483 4837 scope.go:117] "RemoveContainer" containerID="7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0" Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.982964 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:35:04 crc kubenswrapper[4837]: I1001 07:35:04.990026 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n9gsx"] Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.011367 4837 scope.go:117] "RemoveContainer" containerID="b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.040676 4837 scope.go:117] "RemoveContainer" containerID="fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127" Oct 01 07:35:05 crc kubenswrapper[4837]: E1001 07:35:05.041497 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127\": container with ID starting with fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127 not found: ID does not exist" containerID="fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.041554 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127"} err="failed to get container status \"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127\": rpc error: code = NotFound desc = could not find container \"fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127\": container with ID starting with fb7c6275f682f30b189f452a6093cf19e9769f3435803a3543fb2909af714127 not found: ID does not exist" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.041593 4837 scope.go:117] "RemoveContainer" containerID="7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0" Oct 01 07:35:05 crc kubenswrapper[4837]: E1001 07:35:05.042543 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0\": container with ID starting with 7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0 not found: ID does not exist" containerID="7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.042626 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0"} err="failed to get container status \"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0\": rpc error: code = NotFound desc = could not find container \"7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0\": container with ID starting with 7fecf1f429ad0c1fce2b4f4cf2feb571323163541a4b0c45e783a25a894ffcb0 not found: ID does not exist" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.042658 4837 scope.go:117] "RemoveContainer" containerID="b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff" Oct 01 07:35:05 crc kubenswrapper[4837]: E1001 07:35:05.047430 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff\": container with ID starting with b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff not found: ID does not exist" containerID="b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.047497 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff"} err="failed to get container status \"b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff\": rpc error: code = NotFound desc = could not find container \"b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff\": container with ID starting with b8e21cfddee7154ae591c9a9fd63998ae063f67fdd78492c398537d2eb86c6ff not found: ID does not exist" Oct 01 07:35:05 crc kubenswrapper[4837]: I1001 07:35:05.831634 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" path="/var/lib/kubelet/pods/faeb9800-03fb-4c54-9eb3-880238391caf/volumes" Oct 01 07:35:14 crc kubenswrapper[4837]: I1001 07:35:14.517831 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ln2gs" Oct 01 07:35:14 crc kubenswrapper[4837]: I1001 07:35:14.624270 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ln2gs"] Oct 01 07:35:14 crc kubenswrapper[4837]: I1001 07:35:14.723003 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:35:14 crc kubenswrapper[4837]: I1001 07:35:14.723295 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pvrtj" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="registry-server" containerID="cri-o://f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" gracePeriod=2 Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.044445 4837 generic.go:334] "Generic (PLEG): container finished" podID="c9872972-efd5-469f-8908-e4cbe13c2565" containerID="f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" exitCode=0 Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.044754 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerDied","Data":"f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c"} Oct 01 07:35:15 crc kubenswrapper[4837]: E1001 07:35:15.070823 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c is running failed: container process not found" containerID="f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 07:35:15 crc kubenswrapper[4837]: E1001 07:35:15.072924 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c is running failed: container process not found" containerID="f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 07:35:15 crc kubenswrapper[4837]: E1001 07:35:15.075894 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c is running failed: container process not found" containerID="f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" cmd=["grpc_health_probe","-addr=:50051"] Oct 01 07:35:15 crc kubenswrapper[4837]: E1001 07:35:15.075932 4837 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-pvrtj" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="registry-server" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.106675 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.290711 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlv7d\" (UniqueName: \"kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d\") pod \"c9872972-efd5-469f-8908-e4cbe13c2565\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.290801 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities\") pod \"c9872972-efd5-469f-8908-e4cbe13c2565\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.290883 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content\") pod \"c9872972-efd5-469f-8908-e4cbe13c2565\" (UID: \"c9872972-efd5-469f-8908-e4cbe13c2565\") " Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.291365 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities" (OuterVolumeSpecName: "utilities") pod "c9872972-efd5-469f-8908-e4cbe13c2565" (UID: "c9872972-efd5-469f-8908-e4cbe13c2565"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.309632 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d" (OuterVolumeSpecName: "kube-api-access-nlv7d") pod "c9872972-efd5-469f-8908-e4cbe13c2565" (UID: "c9872972-efd5-469f-8908-e4cbe13c2565"). InnerVolumeSpecName "kube-api-access-nlv7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.363052 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9872972-efd5-469f-8908-e4cbe13c2565" (UID: "c9872972-efd5-469f-8908-e4cbe13c2565"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.392264 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.392300 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9872972-efd5-469f-8908-e4cbe13c2565-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.392332 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlv7d\" (UniqueName: \"kubernetes.io/projected/c9872972-efd5-469f-8908-e4cbe13c2565-kube-api-access-nlv7d\") on node \"crc\" DevicePath \"\"" Oct 01 07:35:15 crc kubenswrapper[4837]: I1001 07:35:15.816193 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:35:15 crc kubenswrapper[4837]: E1001 07:35:15.816739 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.060147 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pvrtj" event={"ID":"c9872972-efd5-469f-8908-e4cbe13c2565","Type":"ContainerDied","Data":"5ce507ddb82c0a22eb33d5b90bfcdf5114857a3eeca3d1f0a9a6e74448bef73d"} Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.060239 4837 scope.go:117] "RemoveContainer" containerID="f7df9f8b42fcd598cce01250580164d1891f1a87fa363868aa36d787444bb97c" Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.060314 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pvrtj" Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.093999 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.102972 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pvrtj"] Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.104583 4837 scope.go:117] "RemoveContainer" containerID="3a9f765b114ddae09b09195662e26b966f86c5f6a2d13f9251e811474af89094" Oct 01 07:35:16 crc kubenswrapper[4837]: I1001 07:35:16.138570 4837 scope.go:117] "RemoveContainer" containerID="39cad00979359b607b90a8a63c76d1a0aba83f567bdf78eb02b42c91eb4924fb" Oct 01 07:35:17 crc kubenswrapper[4837]: I1001 07:35:17.833578 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" path="/var/lib/kubelet/pods/c9872972-efd5-469f-8908-e4cbe13c2565/volumes" Oct 01 07:35:29 crc kubenswrapper[4837]: I1001 07:35:29.815750 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:35:29 crc kubenswrapper[4837]: E1001 07:35:29.816506 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:35:40 crc kubenswrapper[4837]: I1001 07:35:40.816536 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:35:40 crc kubenswrapper[4837]: E1001 07:35:40.817761 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:35:55 crc kubenswrapper[4837]: I1001 07:35:55.817129 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:35:56 crc kubenswrapper[4837]: I1001 07:35:56.453826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794"} Oct 01 07:36:03 crc kubenswrapper[4837]: I1001 07:36:03.022116 4837 scope.go:117] "RemoveContainer" containerID="a5007423f20f31799bd3b83421d45c3a51a7e25c84dc7deaaee06262d96de1ef" Oct 01 07:38:23 crc kubenswrapper[4837]: I1001 07:38:23.079942 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:38:23 crc kubenswrapper[4837]: I1001 07:38:23.080543 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:38:53 crc kubenswrapper[4837]: I1001 07:38:53.079683 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:38:53 crc kubenswrapper[4837]: I1001 07:38:53.080341 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.079833 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.080470 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.080540 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.081516 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.081758 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794" gracePeriod=600 Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.404286 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794" exitCode=0 Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.404369 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794"} Oct 01 07:39:23 crc kubenswrapper[4837]: I1001 07:39:23.404662 4837 scope.go:117] "RemoveContainer" containerID="e19cb18ed8de2c74722d93fb6c7d7c748fd1480877802f88174fa4287cb4908d" Oct 01 07:39:24 crc kubenswrapper[4837]: I1001 07:39:24.415685 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399"} Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.884643 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887286 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887329 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887405 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="extract-content" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887426 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="extract-content" Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887457 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887487 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887550 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="extract-utilities" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887571 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="extract-utilities" Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887627 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="extract-utilities" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887643 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="extract-utilities" Oct 01 07:39:30 crc kubenswrapper[4837]: E1001 07:39:30.887672 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="extract-content" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.887688 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="extract-content" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.895023 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9872972-efd5-469f-8908-e4cbe13c2565" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.895095 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="faeb9800-03fb-4c54-9eb3-880238391caf" containerName="registry-server" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.899069 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:30 crc kubenswrapper[4837]: I1001 07:39:30.937035 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.039132 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2frh8\" (UniqueName: \"kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.039183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.039298 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.140458 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2frh8\" (UniqueName: \"kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.140921 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.141062 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.141544 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.141608 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.168405 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2frh8\" (UniqueName: \"kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8\") pod \"redhat-operators-j2stz\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.236064 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:31 crc kubenswrapper[4837]: I1001 07:39:31.695343 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:32 crc kubenswrapper[4837]: I1001 07:39:32.485014 4837 generic.go:334] "Generic (PLEG): container finished" podID="38499749-0014-460a-b0ba-bb5f4cf94911" containerID="b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733" exitCode=0 Oct 01 07:39:32 crc kubenswrapper[4837]: I1001 07:39:32.485113 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerDied","Data":"b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733"} Oct 01 07:39:32 crc kubenswrapper[4837]: I1001 07:39:32.485446 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerStarted","Data":"c4bdc815c23778ea5c6c9e4c8815f09febe56a641cc3cafcf80acc711293a387"} Oct 01 07:39:34 crc kubenswrapper[4837]: I1001 07:39:34.519618 4837 generic.go:334] "Generic (PLEG): container finished" podID="38499749-0014-460a-b0ba-bb5f4cf94911" containerID="60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715" exitCode=0 Oct 01 07:39:34 crc kubenswrapper[4837]: I1001 07:39:34.519921 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerDied","Data":"60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715"} Oct 01 07:39:35 crc kubenswrapper[4837]: I1001 07:39:35.531143 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerStarted","Data":"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178"} Oct 01 07:39:35 crc kubenswrapper[4837]: I1001 07:39:35.564535 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j2stz" podStartSLOduration=2.953947299 podStartE2EDuration="5.56450658s" podCreationTimestamp="2025-10-01 07:39:30 +0000 UTC" firstStartedPulling="2025-10-01 07:39:32.489268472 +0000 UTC m=+2029.330875957" lastFinishedPulling="2025-10-01 07:39:35.099827743 +0000 UTC m=+2031.941435238" observedRunningTime="2025-10-01 07:39:35.557898696 +0000 UTC m=+2032.399506191" watchObservedRunningTime="2025-10-01 07:39:35.56450658 +0000 UTC m=+2032.406114075" Oct 01 07:39:41 crc kubenswrapper[4837]: I1001 07:39:41.236572 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:41 crc kubenswrapper[4837]: I1001 07:39:41.237299 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:41 crc kubenswrapper[4837]: I1001 07:39:41.304198 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:41 crc kubenswrapper[4837]: I1001 07:39:41.698879 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:41 crc kubenswrapper[4837]: I1001 07:39:41.763611 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:43 crc kubenswrapper[4837]: I1001 07:39:43.599948 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j2stz" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="registry-server" containerID="cri-o://5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178" gracePeriod=2 Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.115982 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.260124 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content\") pod \"38499749-0014-460a-b0ba-bb5f4cf94911\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.260197 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities\") pod \"38499749-0014-460a-b0ba-bb5f4cf94911\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.260236 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2frh8\" (UniqueName: \"kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8\") pod \"38499749-0014-460a-b0ba-bb5f4cf94911\" (UID: \"38499749-0014-460a-b0ba-bb5f4cf94911\") " Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.261287 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities" (OuterVolumeSpecName: "utilities") pod "38499749-0014-460a-b0ba-bb5f4cf94911" (UID: "38499749-0014-460a-b0ba-bb5f4cf94911"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.268679 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8" (OuterVolumeSpecName: "kube-api-access-2frh8") pod "38499749-0014-460a-b0ba-bb5f4cf94911" (UID: "38499749-0014-460a-b0ba-bb5f4cf94911"). InnerVolumeSpecName "kube-api-access-2frh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.361671 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.361728 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2frh8\" (UniqueName: \"kubernetes.io/projected/38499749-0014-460a-b0ba-bb5f4cf94911-kube-api-access-2frh8\") on node \"crc\" DevicePath \"\"" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.591068 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38499749-0014-460a-b0ba-bb5f4cf94911" (UID: "38499749-0014-460a-b0ba-bb5f4cf94911"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.613800 4837 generic.go:334] "Generic (PLEG): container finished" podID="38499749-0014-460a-b0ba-bb5f4cf94911" containerID="5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178" exitCode=0 Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.613855 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerDied","Data":"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178"} Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.613895 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j2stz" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.613920 4837 scope.go:117] "RemoveContainer" containerID="5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.613902 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j2stz" event={"ID":"38499749-0014-460a-b0ba-bb5f4cf94911","Type":"ContainerDied","Data":"c4bdc815c23778ea5c6c9e4c8815f09febe56a641cc3cafcf80acc711293a387"} Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.641612 4837 scope.go:117] "RemoveContainer" containerID="60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.654088 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.662307 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j2stz"] Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.666403 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38499749-0014-460a-b0ba-bb5f4cf94911-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.668804 4837 scope.go:117] "RemoveContainer" containerID="b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.700578 4837 scope.go:117] "RemoveContainer" containerID="5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178" Oct 01 07:39:44 crc kubenswrapper[4837]: E1001 07:39:44.701288 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178\": container with ID starting with 5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178 not found: ID does not exist" containerID="5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.701403 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178"} err="failed to get container status \"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178\": rpc error: code = NotFound desc = could not find container \"5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178\": container with ID starting with 5da5a00fbb2588849b484dd0417a3f7afdfbca28d72070ac3846b8b1c033d178 not found: ID does not exist" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.701508 4837 scope.go:117] "RemoveContainer" containerID="60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715" Oct 01 07:39:44 crc kubenswrapper[4837]: E1001 07:39:44.701933 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715\": container with ID starting with 60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715 not found: ID does not exist" containerID="60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.702087 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715"} err="failed to get container status \"60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715\": rpc error: code = NotFound desc = could not find container \"60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715\": container with ID starting with 60ab5348a43499169e6e633018b62648a36ab95ad09e91eda31a7f7df9839715 not found: ID does not exist" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.702200 4837 scope.go:117] "RemoveContainer" containerID="b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733" Oct 01 07:39:44 crc kubenswrapper[4837]: E1001 07:39:44.702593 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733\": container with ID starting with b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733 not found: ID does not exist" containerID="b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733" Oct 01 07:39:44 crc kubenswrapper[4837]: I1001 07:39:44.702732 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733"} err="failed to get container status \"b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733\": rpc error: code = NotFound desc = could not find container \"b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733\": container with ID starting with b3e1f4a72b12d00f6d1df735ffe4f5401c21aecca5a8935a6bf320f9fdf69733 not found: ID does not exist" Oct 01 07:39:45 crc kubenswrapper[4837]: I1001 07:39:45.828275 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" path="/var/lib/kubelet/pods/38499749-0014-460a-b0ba-bb5f4cf94911/volumes" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.080186 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:34 crc kubenswrapper[4837]: E1001 07:40:34.082103 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="extract-utilities" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.082176 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="extract-utilities" Oct 01 07:40:34 crc kubenswrapper[4837]: E1001 07:40:34.082269 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="extract-content" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.082341 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="extract-content" Oct 01 07:40:34 crc kubenswrapper[4837]: E1001 07:40:34.082463 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="registry-server" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.082484 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="registry-server" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.083156 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="38499749-0014-460a-b0ba-bb5f4cf94911" containerName="registry-server" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.087542 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.108675 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.169055 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.169116 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv2r5\" (UniqueName: \"kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.169223 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.270532 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv2r5\" (UniqueName: \"kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.271039 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.271911 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.272099 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.272415 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.292250 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv2r5\" (UniqueName: \"kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5\") pod \"redhat-marketplace-dmh5j\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.422943 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:34 crc kubenswrapper[4837]: I1001 07:40:34.893736 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:35 crc kubenswrapper[4837]: I1001 07:40:35.100352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerStarted","Data":"ca0254a1058bbacd98868a4a145aa9b7f85373cd24092e6de88884c32538e2b7"} Oct 01 07:40:36 crc kubenswrapper[4837]: I1001 07:40:36.113515 4837 generic.go:334] "Generic (PLEG): container finished" podID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerID="42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374" exitCode=0 Oct 01 07:40:36 crc kubenswrapper[4837]: I1001 07:40:36.113625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerDied","Data":"42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374"} Oct 01 07:40:36 crc kubenswrapper[4837]: I1001 07:40:36.116907 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:40:37 crc kubenswrapper[4837]: I1001 07:40:37.128567 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerStarted","Data":"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a"} Oct 01 07:40:38 crc kubenswrapper[4837]: I1001 07:40:38.141514 4837 generic.go:334] "Generic (PLEG): container finished" podID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerID="935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a" exitCode=0 Oct 01 07:40:38 crc kubenswrapper[4837]: I1001 07:40:38.141607 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerDied","Data":"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a"} Oct 01 07:40:39 crc kubenswrapper[4837]: I1001 07:40:39.154659 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerStarted","Data":"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c"} Oct 01 07:40:39 crc kubenswrapper[4837]: I1001 07:40:39.184870 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dmh5j" podStartSLOduration=2.652066855 podStartE2EDuration="5.184834099s" podCreationTimestamp="2025-10-01 07:40:34 +0000 UTC" firstStartedPulling="2025-10-01 07:40:36.116318999 +0000 UTC m=+2092.957926494" lastFinishedPulling="2025-10-01 07:40:38.649086243 +0000 UTC m=+2095.490693738" observedRunningTime="2025-10-01 07:40:39.182982543 +0000 UTC m=+2096.024590028" watchObservedRunningTime="2025-10-01 07:40:39.184834099 +0000 UTC m=+2096.026441594" Oct 01 07:40:44 crc kubenswrapper[4837]: I1001 07:40:44.423896 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:44 crc kubenswrapper[4837]: I1001 07:40:44.424850 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:44 crc kubenswrapper[4837]: I1001 07:40:44.515487 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:45 crc kubenswrapper[4837]: I1001 07:40:45.290106 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:45 crc kubenswrapper[4837]: I1001 07:40:45.357919 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.229152 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dmh5j" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="registry-server" containerID="cri-o://d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c" gracePeriod=2 Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.646462 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.800789 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities\") pod \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.802010 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv2r5\" (UniqueName: \"kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5\") pod \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.802339 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content\") pod \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\" (UID: \"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560\") " Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.802435 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities" (OuterVolumeSpecName: "utilities") pod "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" (UID: "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.803183 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.812147 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5" (OuterVolumeSpecName: "kube-api-access-kv2r5") pod "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" (UID: "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560"). InnerVolumeSpecName "kube-api-access-kv2r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.831288 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" (UID: "ff5e6dd0-f38f-4894-b3f2-869fb7ec8560"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.904767 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:40:47 crc kubenswrapper[4837]: I1001 07:40:47.905136 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv2r5\" (UniqueName: \"kubernetes.io/projected/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560-kube-api-access-kv2r5\") on node \"crc\" DevicePath \"\"" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.242159 4837 generic.go:334] "Generic (PLEG): container finished" podID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerID="d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c" exitCode=0 Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.242206 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerDied","Data":"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c"} Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.242246 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dmh5j" event={"ID":"ff5e6dd0-f38f-4894-b3f2-869fb7ec8560","Type":"ContainerDied","Data":"ca0254a1058bbacd98868a4a145aa9b7f85373cd24092e6de88884c32538e2b7"} Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.242265 4837 scope.go:117] "RemoveContainer" containerID="d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.242184 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dmh5j" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.277095 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.279825 4837 scope.go:117] "RemoveContainer" containerID="935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.289158 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dmh5j"] Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.308344 4837 scope.go:117] "RemoveContainer" containerID="42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.353415 4837 scope.go:117] "RemoveContainer" containerID="d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c" Oct 01 07:40:48 crc kubenswrapper[4837]: E1001 07:40:48.354190 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c\": container with ID starting with d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c not found: ID does not exist" containerID="d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.354252 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c"} err="failed to get container status \"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c\": rpc error: code = NotFound desc = could not find container \"d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c\": container with ID starting with d5a3329a18cbdd87c3515e4e306c9cd6f58fdac16788b56813d9b791112d551c not found: ID does not exist" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.354307 4837 scope.go:117] "RemoveContainer" containerID="935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a" Oct 01 07:40:48 crc kubenswrapper[4837]: E1001 07:40:48.354926 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a\": container with ID starting with 935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a not found: ID does not exist" containerID="935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.354997 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a"} err="failed to get container status \"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a\": rpc error: code = NotFound desc = could not find container \"935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a\": container with ID starting with 935ee5ce3b55d2b3d5876384881837e8614ff72af3a9ee7ef9071fa121dda22a not found: ID does not exist" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.355047 4837 scope.go:117] "RemoveContainer" containerID="42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374" Oct 01 07:40:48 crc kubenswrapper[4837]: E1001 07:40:48.355478 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374\": container with ID starting with 42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374 not found: ID does not exist" containerID="42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374" Oct 01 07:40:48 crc kubenswrapper[4837]: I1001 07:40:48.355544 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374"} err="failed to get container status \"42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374\": rpc error: code = NotFound desc = could not find container \"42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374\": container with ID starting with 42329e0e6e8e98e77a442b14c89bc31c96b695df47746b4c311315544e1d4374 not found: ID does not exist" Oct 01 07:40:49 crc kubenswrapper[4837]: I1001 07:40:49.832191 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" path="/var/lib/kubelet/pods/ff5e6dd0-f38f-4894-b3f2-869fb7ec8560/volumes" Oct 01 07:41:23 crc kubenswrapper[4837]: I1001 07:41:23.080043 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:41:23 crc kubenswrapper[4837]: I1001 07:41:23.080666 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:41:53 crc kubenswrapper[4837]: I1001 07:41:53.079667 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:41:53 crc kubenswrapper[4837]: I1001 07:41:53.080643 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:42:23 crc kubenswrapper[4837]: I1001 07:42:23.080035 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:42:23 crc kubenswrapper[4837]: I1001 07:42:23.080542 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:42:23 crc kubenswrapper[4837]: I1001 07:42:23.080596 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:42:23 crc kubenswrapper[4837]: I1001 07:42:23.081400 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:42:23 crc kubenswrapper[4837]: I1001 07:42:23.081552 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" gracePeriod=600 Oct 01 07:42:23 crc kubenswrapper[4837]: E1001 07:42:23.209679 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:42:24 crc kubenswrapper[4837]: I1001 07:42:24.212242 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" exitCode=0 Oct 01 07:42:24 crc kubenswrapper[4837]: I1001 07:42:24.212314 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399"} Oct 01 07:42:24 crc kubenswrapper[4837]: I1001 07:42:24.212377 4837 scope.go:117] "RemoveContainer" containerID="72411d5e0fb3598faa3efe80f89aa5637c37b0ff5d570f96c7aad7f4a1305794" Oct 01 07:42:24 crc kubenswrapper[4837]: I1001 07:42:24.213240 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:42:24 crc kubenswrapper[4837]: E1001 07:42:24.213798 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:42:37 crc kubenswrapper[4837]: I1001 07:42:37.816781 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:42:37 crc kubenswrapper[4837]: E1001 07:42:37.817926 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:42:50 crc kubenswrapper[4837]: I1001 07:42:50.815938 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:42:50 crc kubenswrapper[4837]: E1001 07:42:50.816939 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:43:05 crc kubenswrapper[4837]: I1001 07:43:05.815544 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:43:05 crc kubenswrapper[4837]: E1001 07:43:05.816237 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:43:17 crc kubenswrapper[4837]: I1001 07:43:17.819473 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:43:17 crc kubenswrapper[4837]: E1001 07:43:17.820184 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:43:28 crc kubenswrapper[4837]: I1001 07:43:28.816196 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:43:28 crc kubenswrapper[4837]: E1001 07:43:28.817000 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:43:43 crc kubenswrapper[4837]: I1001 07:43:43.827207 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:43:43 crc kubenswrapper[4837]: E1001 07:43:43.828819 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:43:55 crc kubenswrapper[4837]: I1001 07:43:55.816163 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:43:55 crc kubenswrapper[4837]: E1001 07:43:55.819089 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:44:10 crc kubenswrapper[4837]: I1001 07:44:10.816158 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:44:10 crc kubenswrapper[4837]: E1001 07:44:10.817231 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:44:25 crc kubenswrapper[4837]: I1001 07:44:25.816728 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:44:25 crc kubenswrapper[4837]: E1001 07:44:25.817783 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:44:36 crc kubenswrapper[4837]: I1001 07:44:36.816110 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:44:36 crc kubenswrapper[4837]: E1001 07:44:36.818668 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:44:50 crc kubenswrapper[4837]: I1001 07:44:50.817165 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:44:50 crc kubenswrapper[4837]: E1001 07:44:50.818155 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.182078 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7"] Oct 01 07:45:00 crc kubenswrapper[4837]: E1001 07:45:00.183986 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="extract-utilities" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.184024 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="extract-utilities" Oct 01 07:45:00 crc kubenswrapper[4837]: E1001 07:45:00.184075 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="registry-server" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.184095 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="registry-server" Oct 01 07:45:00 crc kubenswrapper[4837]: E1001 07:45:00.184128 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="extract-content" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.184149 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="extract-content" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.184614 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5e6dd0-f38f-4894-b3f2-869fb7ec8560" containerName="registry-server" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.188618 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.196614 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.196967 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.213984 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7"] Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.286410 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.286485 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.286549 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htxl6\" (UniqueName: \"kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.387888 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.387958 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.388032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htxl6\" (UniqueName: \"kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.389090 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.393374 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.416168 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htxl6\" (UniqueName: \"kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6\") pod \"collect-profiles-29321745-9h6l7\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:00 crc kubenswrapper[4837]: I1001 07:45:00.520934 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:01 crc kubenswrapper[4837]: I1001 07:45:01.006161 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7"] Oct 01 07:45:01 crc kubenswrapper[4837]: I1001 07:45:01.693901 4837 generic.go:334] "Generic (PLEG): container finished" podID="ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" containerID="4f4286aa037cb6b9afe186444d61e6a9d5a3792738d6df3eea1f39d79cb705dc" exitCode=0 Oct 01 07:45:01 crc kubenswrapper[4837]: I1001 07:45:01.694024 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" event={"ID":"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46","Type":"ContainerDied","Data":"4f4286aa037cb6b9afe186444d61e6a9d5a3792738d6df3eea1f39d79cb705dc"} Oct 01 07:45:01 crc kubenswrapper[4837]: I1001 07:45:01.694586 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" event={"ID":"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46","Type":"ContainerStarted","Data":"aea3bb26c3d5fd8da60af6329d7975f8ec8d01b0a8e40db1ba924bff7571806d"} Oct 01 07:45:01 crc kubenswrapper[4837]: I1001 07:45:01.816872 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:45:01 crc kubenswrapper[4837]: E1001 07:45:01.817409 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.054051 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.133207 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htxl6\" (UniqueName: \"kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6\") pod \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.133324 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume\") pod \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.133454 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume\") pod \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\" (UID: \"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46\") " Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.134906 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume" (OuterVolumeSpecName: "config-volume") pod "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" (UID: "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.143507 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6" (OuterVolumeSpecName: "kube-api-access-htxl6") pod "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" (UID: "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46"). InnerVolumeSpecName "kube-api-access-htxl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.143580 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" (UID: "ffc6a77a-70f6-4100-9197-a5e0aa1c4d46"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.236556 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.236618 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.236640 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htxl6\" (UniqueName: \"kubernetes.io/projected/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46-kube-api-access-htxl6\") on node \"crc\" DevicePath \"\"" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.716076 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" event={"ID":"ffc6a77a-70f6-4100-9197-a5e0aa1c4d46","Type":"ContainerDied","Data":"aea3bb26c3d5fd8da60af6329d7975f8ec8d01b0a8e40db1ba924bff7571806d"} Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.716126 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aea3bb26c3d5fd8da60af6329d7975f8ec8d01b0a8e40db1ba924bff7571806d" Oct 01 07:45:03 crc kubenswrapper[4837]: I1001 07:45:03.716207 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7" Oct 01 07:45:04 crc kubenswrapper[4837]: I1001 07:45:04.153331 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l"] Oct 01 07:45:04 crc kubenswrapper[4837]: I1001 07:45:04.160905 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321700-vhv2l"] Oct 01 07:45:05 crc kubenswrapper[4837]: I1001 07:45:05.827510 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117" path="/var/lib/kubelet/pods/b7fc4d5e-2b1f-4db6-8e0c-d261dc3df117/volumes" Oct 01 07:45:12 crc kubenswrapper[4837]: I1001 07:45:12.816595 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:45:12 crc kubenswrapper[4837]: E1001 07:45:12.817415 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:45:25 crc kubenswrapper[4837]: I1001 07:45:25.815866 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:45:25 crc kubenswrapper[4837]: E1001 07:45:25.817104 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:45:38 crc kubenswrapper[4837]: I1001 07:45:38.815852 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:45:38 crc kubenswrapper[4837]: E1001 07:45:38.816799 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:45:51 crc kubenswrapper[4837]: I1001 07:45:51.816948 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:45:51 crc kubenswrapper[4837]: E1001 07:45:51.818531 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:46:03 crc kubenswrapper[4837]: I1001 07:46:03.320729 4837 scope.go:117] "RemoveContainer" containerID="4eaa95cefbf0e93e9c4e1eeb236af703fb15e897d88aa2c8168363eb2e94af09" Oct 01 07:46:04 crc kubenswrapper[4837]: I1001 07:46:04.817732 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:46:04 crc kubenswrapper[4837]: E1001 07:46:04.818600 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:46:15 crc kubenswrapper[4837]: I1001 07:46:15.815808 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:46:15 crc kubenswrapper[4837]: E1001 07:46:15.816632 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:46:27 crc kubenswrapper[4837]: I1001 07:46:27.816870 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:46:27 crc kubenswrapper[4837]: E1001 07:46:27.817617 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:46:41 crc kubenswrapper[4837]: I1001 07:46:41.815995 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:46:41 crc kubenswrapper[4837]: E1001 07:46:41.817082 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:46:52 crc kubenswrapper[4837]: I1001 07:46:52.816172 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:46:52 crc kubenswrapper[4837]: E1001 07:46:52.816943 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:47:03 crc kubenswrapper[4837]: I1001 07:47:03.824202 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:47:03 crc kubenswrapper[4837]: E1001 07:47:03.825167 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:47:17 crc kubenswrapper[4837]: I1001 07:47:17.816135 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:47:17 crc kubenswrapper[4837]: E1001 07:47:17.816751 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:47:31 crc kubenswrapper[4837]: I1001 07:47:31.816183 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:47:34 crc kubenswrapper[4837]: I1001 07:47:34.086210 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc"} Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.221599 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:11 crc kubenswrapper[4837]: E1001 07:48:11.223075 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" containerName="collect-profiles" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.223095 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" containerName="collect-profiles" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.223340 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" containerName="collect-profiles" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.225003 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.250712 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.402949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.403018 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbrmw\" (UniqueName: \"kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.403039 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.504942 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.505057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbrmw\" (UniqueName: \"kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.505103 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.505777 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.505843 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.542280 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbrmw\" (UniqueName: \"kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw\") pod \"certified-operators-cg9m6\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:11 crc kubenswrapper[4837]: I1001 07:48:11.560256 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.075725 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.421043 4837 generic.go:334] "Generic (PLEG): container finished" podID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerID="ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5" exitCode=0 Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.421106 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerDied","Data":"ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5"} Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.421139 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerStarted","Data":"33e2c7be5384d9e60d08a84b2ff45285a246a17625a44955eca8f0580e73e13b"} Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.423931 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.619640 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.621330 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.637229 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.728212 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szdjl\" (UniqueName: \"kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.728258 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.728466 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.829548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.829668 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szdjl\" (UniqueName: \"kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.829982 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.830269 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.831003 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.853623 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szdjl\" (UniqueName: \"kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl\") pod \"community-operators-pzgbm\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:12 crc kubenswrapper[4837]: I1001 07:48:12.943576 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:13 crc kubenswrapper[4837]: I1001 07:48:13.527173 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:14 crc kubenswrapper[4837]: I1001 07:48:14.438894 4837 generic.go:334] "Generic (PLEG): container finished" podID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerID="394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606" exitCode=0 Oct 01 07:48:14 crc kubenswrapper[4837]: I1001 07:48:14.439006 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerDied","Data":"394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606"} Oct 01 07:48:14 crc kubenswrapper[4837]: I1001 07:48:14.439619 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerStarted","Data":"1699c9fef62717df9be7033fcc12bdff82968639871e59333ca13513123da19b"} Oct 01 07:48:14 crc kubenswrapper[4837]: I1001 07:48:14.446328 4837 generic.go:334] "Generic (PLEG): container finished" podID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerID="afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0" exitCode=0 Oct 01 07:48:14 crc kubenswrapper[4837]: I1001 07:48:14.446393 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerDied","Data":"afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0"} Oct 01 07:48:15 crc kubenswrapper[4837]: I1001 07:48:15.457152 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerStarted","Data":"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b"} Oct 01 07:48:15 crc kubenswrapper[4837]: I1001 07:48:15.487634 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cg9m6" podStartSLOduration=2.025822662 podStartE2EDuration="4.487606031s" podCreationTimestamp="2025-10-01 07:48:11 +0000 UTC" firstStartedPulling="2025-10-01 07:48:12.423338581 +0000 UTC m=+2549.264946076" lastFinishedPulling="2025-10-01 07:48:14.88512199 +0000 UTC m=+2551.726729445" observedRunningTime="2025-10-01 07:48:15.477041991 +0000 UTC m=+2552.318649476" watchObservedRunningTime="2025-10-01 07:48:15.487606031 +0000 UTC m=+2552.329213516" Oct 01 07:48:16 crc kubenswrapper[4837]: I1001 07:48:16.469252 4837 generic.go:334] "Generic (PLEG): container finished" podID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerID="490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2" exitCode=0 Oct 01 07:48:16 crc kubenswrapper[4837]: I1001 07:48:16.469346 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerDied","Data":"490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2"} Oct 01 07:48:17 crc kubenswrapper[4837]: I1001 07:48:17.478927 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerStarted","Data":"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f"} Oct 01 07:48:17 crc kubenswrapper[4837]: I1001 07:48:17.504884 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pzgbm" podStartSLOduration=3.014249535 podStartE2EDuration="5.504854642s" podCreationTimestamp="2025-10-01 07:48:12 +0000 UTC" firstStartedPulling="2025-10-01 07:48:14.443560405 +0000 UTC m=+2551.285167870" lastFinishedPulling="2025-10-01 07:48:16.934165482 +0000 UTC m=+2553.775772977" observedRunningTime="2025-10-01 07:48:17.497149372 +0000 UTC m=+2554.338756847" watchObservedRunningTime="2025-10-01 07:48:17.504854642 +0000 UTC m=+2554.346462127" Oct 01 07:48:21 crc kubenswrapper[4837]: I1001 07:48:21.560608 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:21 crc kubenswrapper[4837]: I1001 07:48:21.561241 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:21 crc kubenswrapper[4837]: I1001 07:48:21.660275 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:22 crc kubenswrapper[4837]: I1001 07:48:22.600221 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:22 crc kubenswrapper[4837]: I1001 07:48:22.679910 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:22 crc kubenswrapper[4837]: I1001 07:48:22.944934 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:22 crc kubenswrapper[4837]: I1001 07:48:22.945011 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:23 crc kubenswrapper[4837]: I1001 07:48:23.025013 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:23 crc kubenswrapper[4837]: I1001 07:48:23.612783 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:24 crc kubenswrapper[4837]: I1001 07:48:24.546132 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cg9m6" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="registry-server" containerID="cri-o://afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b" gracePeriod=2 Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.013571 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.538439 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.561432 4837 generic.go:334] "Generic (PLEG): container finished" podID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerID="afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b" exitCode=0 Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.561491 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cg9m6" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.561498 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerDied","Data":"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b"} Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.561606 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cg9m6" event={"ID":"e8e549ce-cc53-4ab2-883e-1229ca5272bc","Type":"ContainerDied","Data":"33e2c7be5384d9e60d08a84b2ff45285a246a17625a44955eca8f0580e73e13b"} Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.561634 4837 scope.go:117] "RemoveContainer" containerID="afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.562004 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pzgbm" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="registry-server" containerID="cri-o://ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f" gracePeriod=2 Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.588548 4837 scope.go:117] "RemoveContainer" containerID="afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.625051 4837 scope.go:117] "RemoveContainer" containerID="ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.637412 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbrmw\" (UniqueName: \"kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw\") pod \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.637501 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content\") pod \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.637532 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities\") pod \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\" (UID: \"e8e549ce-cc53-4ab2-883e-1229ca5272bc\") " Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.639145 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities" (OuterVolumeSpecName: "utilities") pod "e8e549ce-cc53-4ab2-883e-1229ca5272bc" (UID: "e8e549ce-cc53-4ab2-883e-1229ca5272bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.646057 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw" (OuterVolumeSpecName: "kube-api-access-dbrmw") pod "e8e549ce-cc53-4ab2-883e-1229ca5272bc" (UID: "e8e549ce-cc53-4ab2-883e-1229ca5272bc"). InnerVolumeSpecName "kube-api-access-dbrmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.692051 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8e549ce-cc53-4ab2-883e-1229ca5272bc" (UID: "e8e549ce-cc53-4ab2-883e-1229ca5272bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.739405 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.739435 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8e549ce-cc53-4ab2-883e-1229ca5272bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.739554 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbrmw\" (UniqueName: \"kubernetes.io/projected/e8e549ce-cc53-4ab2-883e-1229ca5272bc-kube-api-access-dbrmw\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.753796 4837 scope.go:117] "RemoveContainer" containerID="afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b" Oct 01 07:48:25 crc kubenswrapper[4837]: E1001 07:48:25.754308 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b\": container with ID starting with afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b not found: ID does not exist" containerID="afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.754353 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b"} err="failed to get container status \"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b\": rpc error: code = NotFound desc = could not find container \"afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b\": container with ID starting with afeb2b0f0ff34cdbcbbfc4a5c708e4d2ac94ee354a8ecc8e0591997d88f8d25b not found: ID does not exist" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.754379 4837 scope.go:117] "RemoveContainer" containerID="afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0" Oct 01 07:48:25 crc kubenswrapper[4837]: E1001 07:48:25.755333 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0\": container with ID starting with afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0 not found: ID does not exist" containerID="afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.755362 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0"} err="failed to get container status \"afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0\": rpc error: code = NotFound desc = could not find container \"afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0\": container with ID starting with afe3d225922a648eac0b39025584813f0a3b8914b6cda36a073c417ea2f462b0 not found: ID does not exist" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.755379 4837 scope.go:117] "RemoveContainer" containerID="ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5" Oct 01 07:48:25 crc kubenswrapper[4837]: E1001 07:48:25.755847 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5\": container with ID starting with ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5 not found: ID does not exist" containerID="ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.755915 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5"} err="failed to get container status \"ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5\": rpc error: code = NotFound desc = could not find container \"ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5\": container with ID starting with ae6ac152ca8f7ecce717e35724eeff69fb93e7e95511ed076ccb707a7dea0de5 not found: ID does not exist" Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.877529 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.896850 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cg9m6"] Oct 01 07:48:25 crc kubenswrapper[4837]: I1001 07:48:25.985188 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.044060 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities\") pod \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.044230 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content\") pod \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.044365 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szdjl\" (UniqueName: \"kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl\") pod \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\" (UID: \"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c\") " Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.044891 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities" (OuterVolumeSpecName: "utilities") pod "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" (UID: "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.048011 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl" (OuterVolumeSpecName: "kube-api-access-szdjl") pod "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" (UID: "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c"). InnerVolumeSpecName "kube-api-access-szdjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.100226 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" (UID: "ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.146575 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.146609 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szdjl\" (UniqueName: \"kubernetes.io/projected/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-kube-api-access-szdjl\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.146623 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.573950 4837 generic.go:334] "Generic (PLEG): container finished" podID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerID="ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f" exitCode=0 Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.574034 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerDied","Data":"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f"} Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.574108 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pzgbm" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.574378 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pzgbm" event={"ID":"ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c","Type":"ContainerDied","Data":"1699c9fef62717df9be7033fcc12bdff82968639871e59333ca13513123da19b"} Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.574415 4837 scope.go:117] "RemoveContainer" containerID="ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.601646 4837 scope.go:117] "RemoveContainer" containerID="490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.633507 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.643075 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pzgbm"] Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.644031 4837 scope.go:117] "RemoveContainer" containerID="394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.675991 4837 scope.go:117] "RemoveContainer" containerID="ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f" Oct 01 07:48:26 crc kubenswrapper[4837]: E1001 07:48:26.676487 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f\": container with ID starting with ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f not found: ID does not exist" containerID="ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.676531 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f"} err="failed to get container status \"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f\": rpc error: code = NotFound desc = could not find container \"ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f\": container with ID starting with ce8f3906900be69412b78ae67c84c26765b5f0710725be5170d6e27f8aefd32f not found: ID does not exist" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.676557 4837 scope.go:117] "RemoveContainer" containerID="490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2" Oct 01 07:48:26 crc kubenswrapper[4837]: E1001 07:48:26.677159 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2\": container with ID starting with 490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2 not found: ID does not exist" containerID="490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.677232 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2"} err="failed to get container status \"490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2\": rpc error: code = NotFound desc = could not find container \"490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2\": container with ID starting with 490fbb5f8fe2805e012335c4bbff48cb3e39a81d43b523e538b57d57cd5830c2 not found: ID does not exist" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.677286 4837 scope.go:117] "RemoveContainer" containerID="394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606" Oct 01 07:48:26 crc kubenswrapper[4837]: E1001 07:48:26.677809 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606\": container with ID starting with 394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606 not found: ID does not exist" containerID="394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606" Oct 01 07:48:26 crc kubenswrapper[4837]: I1001 07:48:26.677839 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606"} err="failed to get container status \"394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606\": rpc error: code = NotFound desc = could not find container \"394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606\": container with ID starting with 394bb0cd66e809888a4b967c77347d53ffa295640770310ac381a30b921c0606 not found: ID does not exist" Oct 01 07:48:27 crc kubenswrapper[4837]: I1001 07:48:27.832814 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" path="/var/lib/kubelet/pods/ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c/volumes" Oct 01 07:48:27 crc kubenswrapper[4837]: I1001 07:48:27.834366 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" path="/var/lib/kubelet/pods/e8e549ce-cc53-4ab2-883e-1229ca5272bc/volumes" Oct 01 07:49:53 crc kubenswrapper[4837]: I1001 07:49:53.080100 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:49:53 crc kubenswrapper[4837]: I1001 07:49:53.082674 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:50:23 crc kubenswrapper[4837]: I1001 07:50:23.079739 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:50:23 crc kubenswrapper[4837]: I1001 07:50:23.082802 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.050178 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.050890 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="extract-utilities" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.050924 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="extract-utilities" Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.050960 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="extract-utilities" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.050975 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="extract-utilities" Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.051006 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="extract-content" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051020 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="extract-content" Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.051040 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="extract-content" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051053 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="extract-content" Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.051076 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051088 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: E1001 07:50:24.051114 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051126 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051370 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4bf57b-3b1c-4cf4-9754-fe12b7d4764c" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.051407 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8e549ce-cc53-4ab2-883e-1229ca5272bc" containerName="registry-server" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.053265 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.075503 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.164293 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.164394 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.164461 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch6br\" (UniqueName: \"kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.265757 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch6br\" (UniqueName: \"kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.265861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.265920 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.266579 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.266769 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.290387 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch6br\" (UniqueName: \"kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br\") pod \"redhat-operators-phm28\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.429134 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:24 crc kubenswrapper[4837]: I1001 07:50:24.935349 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:25 crc kubenswrapper[4837]: I1001 07:50:25.791496 4837 generic.go:334] "Generic (PLEG): container finished" podID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerID="b893f6e1d0f9cfb51b88eada725e9f6434a421ba6c14dde032f5cf1bf8d91639" exitCode=0 Oct 01 07:50:25 crc kubenswrapper[4837]: I1001 07:50:25.791607 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerDied","Data":"b893f6e1d0f9cfb51b88eada725e9f6434a421ba6c14dde032f5cf1bf8d91639"} Oct 01 07:50:25 crc kubenswrapper[4837]: I1001 07:50:25.792010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerStarted","Data":"5aad2f453f10ba12191be7536c636a7e387b5002bf1b2765284932c77c121c63"} Oct 01 07:50:27 crc kubenswrapper[4837]: I1001 07:50:27.808841 4837 generic.go:334] "Generic (PLEG): container finished" podID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerID="f3892f75a78fe5c840d40f1dabaf033e81293f18f56326f7a84cedc8476279f9" exitCode=0 Oct 01 07:50:27 crc kubenswrapper[4837]: I1001 07:50:27.808887 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerDied","Data":"f3892f75a78fe5c840d40f1dabaf033e81293f18f56326f7a84cedc8476279f9"} Oct 01 07:50:27 crc kubenswrapper[4837]: E1001 07:50:27.908138 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf563afa7_3ca8_4c09_b0ac_030be2ce50f7.slice/crio-f3892f75a78fe5c840d40f1dabaf033e81293f18f56326f7a84cedc8476279f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf563afa7_3ca8_4c09_b0ac_030be2ce50f7.slice/crio-conmon-f3892f75a78fe5c840d40f1dabaf033e81293f18f56326f7a84cedc8476279f9.scope\": RecentStats: unable to find data in memory cache]" Oct 01 07:50:28 crc kubenswrapper[4837]: I1001 07:50:28.823149 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerStarted","Data":"7f87654a7d048faa9330eb52c1c497547102e9a7d8c02013f118b333e40588f7"} Oct 01 07:50:28 crc kubenswrapper[4837]: I1001 07:50:28.852731 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-phm28" podStartSLOduration=2.2385468570000002 podStartE2EDuration="4.852685821s" podCreationTimestamp="2025-10-01 07:50:24 +0000 UTC" firstStartedPulling="2025-10-01 07:50:25.795268679 +0000 UTC m=+2682.636876174" lastFinishedPulling="2025-10-01 07:50:28.409407643 +0000 UTC m=+2685.251015138" observedRunningTime="2025-10-01 07:50:28.848203061 +0000 UTC m=+2685.689810556" watchObservedRunningTime="2025-10-01 07:50:28.852685821 +0000 UTC m=+2685.694293306" Oct 01 07:50:34 crc kubenswrapper[4837]: I1001 07:50:34.429378 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:34 crc kubenswrapper[4837]: I1001 07:50:34.431416 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:35 crc kubenswrapper[4837]: I1001 07:50:35.506933 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-phm28" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="registry-server" probeResult="failure" output=< Oct 01 07:50:35 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 07:50:35 crc kubenswrapper[4837]: > Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.613803 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.617675 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.637858 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.743933 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.744066 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh5fs\" (UniqueName: \"kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.744135 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.845785 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.845906 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh5fs\" (UniqueName: \"kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.845991 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.846530 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.846528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.870644 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh5fs\" (UniqueName: \"kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs\") pod \"redhat-marketplace-jnnp5\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:41 crc kubenswrapper[4837]: I1001 07:50:41.952473 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:42 crc kubenswrapper[4837]: I1001 07:50:42.542665 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:42 crc kubenswrapper[4837]: I1001 07:50:42.949747 4837 generic.go:334] "Generic (PLEG): container finished" podID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerID="ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f" exitCode=0 Oct 01 07:50:42 crc kubenswrapper[4837]: I1001 07:50:42.949831 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerDied","Data":"ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f"} Oct 01 07:50:42 crc kubenswrapper[4837]: I1001 07:50:42.951466 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerStarted","Data":"f15afd6275ae3661374f22af109d5af9152eb2d64bca4603c29c0d1196a2ee5a"} Oct 01 07:50:43 crc kubenswrapper[4837]: I1001 07:50:43.979801 4837 generic.go:334] "Generic (PLEG): container finished" podID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerID="0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108" exitCode=0 Oct 01 07:50:43 crc kubenswrapper[4837]: I1001 07:50:43.980228 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerDied","Data":"0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108"} Oct 01 07:50:44 crc kubenswrapper[4837]: I1001 07:50:44.501921 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:44 crc kubenswrapper[4837]: I1001 07:50:44.564346 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:44 crc kubenswrapper[4837]: I1001 07:50:44.993803 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerStarted","Data":"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d"} Oct 01 07:50:45 crc kubenswrapper[4837]: I1001 07:50:45.026191 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jnnp5" podStartSLOduration=2.293643545 podStartE2EDuration="4.026169846s" podCreationTimestamp="2025-10-01 07:50:41 +0000 UTC" firstStartedPulling="2025-10-01 07:50:42.952410307 +0000 UTC m=+2699.794017762" lastFinishedPulling="2025-10-01 07:50:44.684936578 +0000 UTC m=+2701.526544063" observedRunningTime="2025-10-01 07:50:45.019208645 +0000 UTC m=+2701.860816110" watchObservedRunningTime="2025-10-01 07:50:45.026169846 +0000 UTC m=+2701.867777311" Oct 01 07:50:46 crc kubenswrapper[4837]: I1001 07:50:46.769219 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:46 crc kubenswrapper[4837]: I1001 07:50:46.770157 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-phm28" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="registry-server" containerID="cri-o://7f87654a7d048faa9330eb52c1c497547102e9a7d8c02013f118b333e40588f7" gracePeriod=2 Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.016285 4837 generic.go:334] "Generic (PLEG): container finished" podID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerID="7f87654a7d048faa9330eb52c1c497547102e9a7d8c02013f118b333e40588f7" exitCode=0 Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.016367 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerDied","Data":"7f87654a7d048faa9330eb52c1c497547102e9a7d8c02013f118b333e40588f7"} Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.240879 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.326215 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content\") pod \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.326325 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch6br\" (UniqueName: \"kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br\") pod \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.326359 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities\") pod \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\" (UID: \"f563afa7-3ca8-4c09-b0ac-030be2ce50f7\") " Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.327787 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities" (OuterVolumeSpecName: "utilities") pod "f563afa7-3ca8-4c09-b0ac-030be2ce50f7" (UID: "f563afa7-3ca8-4c09-b0ac-030be2ce50f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.334459 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br" (OuterVolumeSpecName: "kube-api-access-ch6br") pod "f563afa7-3ca8-4c09-b0ac-030be2ce50f7" (UID: "f563afa7-3ca8-4c09-b0ac-030be2ce50f7"). InnerVolumeSpecName "kube-api-access-ch6br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.428406 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch6br\" (UniqueName: \"kubernetes.io/projected/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-kube-api-access-ch6br\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.428437 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.437174 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f563afa7-3ca8-4c09-b0ac-030be2ce50f7" (UID: "f563afa7-3ca8-4c09-b0ac-030be2ce50f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:50:47 crc kubenswrapper[4837]: I1001 07:50:47.530302 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f563afa7-3ca8-4c09-b0ac-030be2ce50f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.030475 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phm28" event={"ID":"f563afa7-3ca8-4c09-b0ac-030be2ce50f7","Type":"ContainerDied","Data":"5aad2f453f10ba12191be7536c636a7e387b5002bf1b2765284932c77c121c63"} Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.030577 4837 scope.go:117] "RemoveContainer" containerID="7f87654a7d048faa9330eb52c1c497547102e9a7d8c02013f118b333e40588f7" Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.030581 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phm28" Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.063857 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.066091 4837 scope.go:117] "RemoveContainer" containerID="f3892f75a78fe5c840d40f1dabaf033e81293f18f56326f7a84cedc8476279f9" Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.074014 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-phm28"] Oct 01 07:50:48 crc kubenswrapper[4837]: I1001 07:50:48.099074 4837 scope.go:117] "RemoveContainer" containerID="b893f6e1d0f9cfb51b88eada725e9f6434a421ba6c14dde032f5cf1bf8d91639" Oct 01 07:50:49 crc kubenswrapper[4837]: I1001 07:50:49.828682 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" path="/var/lib/kubelet/pods/f563afa7-3ca8-4c09-b0ac-030be2ce50f7/volumes" Oct 01 07:50:51 crc kubenswrapper[4837]: I1001 07:50:51.953668 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:51 crc kubenswrapper[4837]: I1001 07:50:51.954371 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:52 crc kubenswrapper[4837]: I1001 07:50:52.037987 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:52 crc kubenswrapper[4837]: I1001 07:50:52.140052 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:52 crc kubenswrapper[4837]: I1001 07:50:52.290732 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:53 crc kubenswrapper[4837]: I1001 07:50:53.079278 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:50:53 crc kubenswrapper[4837]: I1001 07:50:53.079378 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:50:53 crc kubenswrapper[4837]: I1001 07:50:53.079458 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:50:53 crc kubenswrapper[4837]: I1001 07:50:53.080620 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:50:53 crc kubenswrapper[4837]: I1001 07:50:53.080772 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc" gracePeriod=600 Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.095930 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc" exitCode=0 Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.095982 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc"} Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.096568 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326"} Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.096600 4837 scope.go:117] "RemoveContainer" containerID="5a3104524dda32b55ee583c3e13358de841fdc00158ea64219bbb6e4edf10399" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.096804 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jnnp5" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="registry-server" containerID="cri-o://a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d" gracePeriod=2 Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.568752 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.656789 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities\") pod \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.656891 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh5fs\" (UniqueName: \"kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs\") pod \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.656930 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content\") pod \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\" (UID: \"ad3754cd-3b91-48d6-be0d-8af12e159cb1\") " Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.657786 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities" (OuterVolumeSpecName: "utilities") pod "ad3754cd-3b91-48d6-be0d-8af12e159cb1" (UID: "ad3754cd-3b91-48d6-be0d-8af12e159cb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.669420 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad3754cd-3b91-48d6-be0d-8af12e159cb1" (UID: "ad3754cd-3b91-48d6-be0d-8af12e159cb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.669617 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs" (OuterVolumeSpecName: "kube-api-access-qh5fs") pod "ad3754cd-3b91-48d6-be0d-8af12e159cb1" (UID: "ad3754cd-3b91-48d6-be0d-8af12e159cb1"). InnerVolumeSpecName "kube-api-access-qh5fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.758857 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.758898 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad3754cd-3b91-48d6-be0d-8af12e159cb1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:54 crc kubenswrapper[4837]: I1001 07:50:54.758912 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh5fs\" (UniqueName: \"kubernetes.io/projected/ad3754cd-3b91-48d6-be0d-8af12e159cb1-kube-api-access-qh5fs\") on node \"crc\" DevicePath \"\"" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.110270 4837 generic.go:334] "Generic (PLEG): container finished" podID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerID="a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d" exitCode=0 Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.110364 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jnnp5" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.110386 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerDied","Data":"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d"} Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.110832 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jnnp5" event={"ID":"ad3754cd-3b91-48d6-be0d-8af12e159cb1","Type":"ContainerDied","Data":"f15afd6275ae3661374f22af109d5af9152eb2d64bca4603c29c0d1196a2ee5a"} Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.110865 4837 scope.go:117] "RemoveContainer" containerID="a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.139397 4837 scope.go:117] "RemoveContainer" containerID="0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.166533 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.174952 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jnnp5"] Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.180375 4837 scope.go:117] "RemoveContainer" containerID="ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.224117 4837 scope.go:117] "RemoveContainer" containerID="a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d" Oct 01 07:50:55 crc kubenswrapper[4837]: E1001 07:50:55.224603 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d\": container with ID starting with a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d not found: ID does not exist" containerID="a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.224668 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d"} err="failed to get container status \"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d\": rpc error: code = NotFound desc = could not find container \"a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d\": container with ID starting with a43bf997461da155c9b780e7e8e8239ff68ce01922ffa64a0e8d41a863e7a82d not found: ID does not exist" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.224737 4837 scope.go:117] "RemoveContainer" containerID="0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108" Oct 01 07:50:55 crc kubenswrapper[4837]: E1001 07:50:55.225187 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108\": container with ID starting with 0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108 not found: ID does not exist" containerID="0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.225250 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108"} err="failed to get container status \"0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108\": rpc error: code = NotFound desc = could not find container \"0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108\": container with ID starting with 0a46197bceaa302928b703aca017be95e612be641406460ce7582866ca267108 not found: ID does not exist" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.225296 4837 scope.go:117] "RemoveContainer" containerID="ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f" Oct 01 07:50:55 crc kubenswrapper[4837]: E1001 07:50:55.225851 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f\": container with ID starting with ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f not found: ID does not exist" containerID="ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.225902 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f"} err="failed to get container status \"ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f\": rpc error: code = NotFound desc = could not find container \"ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f\": container with ID starting with ac152cee79f1212746843dfb3a518fd75a73b12692f3bfaec7b0bb51ed303c4f not found: ID does not exist" Oct 01 07:50:55 crc kubenswrapper[4837]: I1001 07:50:55.831065 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" path="/var/lib/kubelet/pods/ad3754cd-3b91-48d6-be0d-8af12e159cb1/volumes" Oct 01 07:52:53 crc kubenswrapper[4837]: I1001 07:52:53.079613 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:52:53 crc kubenswrapper[4837]: I1001 07:52:53.080293 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:53:23 crc kubenswrapper[4837]: I1001 07:53:23.080310 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:53:23 crc kubenswrapper[4837]: I1001 07:53:23.081443 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.079554 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.080349 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.080418 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.081436 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.081534 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" gracePeriod=600 Oct 01 07:53:53 crc kubenswrapper[4837]: E1001 07:53:53.216058 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.824779 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" exitCode=0 Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.824828 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326"} Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.824858 4837 scope.go:117] "RemoveContainer" containerID="7f14f7374a8386a4ee12359eda9de2adfbf66f670ee2d12c4ab2f2adccf772dc" Oct 01 07:53:53 crc kubenswrapper[4837]: I1001 07:53:53.825594 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:53:53 crc kubenswrapper[4837]: E1001 07:53:53.826020 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:54:04 crc kubenswrapper[4837]: I1001 07:54:04.816419 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:54:04 crc kubenswrapper[4837]: E1001 07:54:04.817398 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:54:17 crc kubenswrapper[4837]: I1001 07:54:17.816526 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:54:17 crc kubenswrapper[4837]: E1001 07:54:17.817968 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:54:32 crc kubenswrapper[4837]: I1001 07:54:32.815920 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:54:32 crc kubenswrapper[4837]: E1001 07:54:32.816530 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:54:44 crc kubenswrapper[4837]: I1001 07:54:44.816829 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:54:44 crc kubenswrapper[4837]: E1001 07:54:44.817784 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:54:59 crc kubenswrapper[4837]: I1001 07:54:59.816413 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:54:59 crc kubenswrapper[4837]: E1001 07:54:59.817930 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:55:10 crc kubenswrapper[4837]: I1001 07:55:10.816389 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:55:10 crc kubenswrapper[4837]: E1001 07:55:10.817275 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:55:24 crc kubenswrapper[4837]: I1001 07:55:24.816811 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:55:24 crc kubenswrapper[4837]: E1001 07:55:24.817723 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:55:37 crc kubenswrapper[4837]: I1001 07:55:37.816573 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:55:37 crc kubenswrapper[4837]: E1001 07:55:37.819203 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:55:48 crc kubenswrapper[4837]: I1001 07:55:48.816098 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:55:48 crc kubenswrapper[4837]: E1001 07:55:48.817207 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:56:00 crc kubenswrapper[4837]: I1001 07:56:00.816156 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:56:00 crc kubenswrapper[4837]: E1001 07:56:00.816866 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:56:13 crc kubenswrapper[4837]: I1001 07:56:13.824359 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:56:13 crc kubenswrapper[4837]: E1001 07:56:13.826531 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:56:25 crc kubenswrapper[4837]: I1001 07:56:25.816932 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:56:25 crc kubenswrapper[4837]: E1001 07:56:25.818672 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:56:38 crc kubenswrapper[4837]: I1001 07:56:38.816356 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:56:38 crc kubenswrapper[4837]: E1001 07:56:38.817502 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:56:52 crc kubenswrapper[4837]: I1001 07:56:52.816501 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:56:52 crc kubenswrapper[4837]: E1001 07:56:52.817886 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:57:04 crc kubenswrapper[4837]: I1001 07:57:04.817900 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:57:04 crc kubenswrapper[4837]: E1001 07:57:04.819046 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:57:15 crc kubenswrapper[4837]: I1001 07:57:15.816662 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:57:15 crc kubenswrapper[4837]: E1001 07:57:15.818068 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:57:27 crc kubenswrapper[4837]: I1001 07:57:27.816754 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:57:27 crc kubenswrapper[4837]: E1001 07:57:27.817898 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:57:40 crc kubenswrapper[4837]: I1001 07:57:40.816894 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:57:40 crc kubenswrapper[4837]: E1001 07:57:40.818049 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:57:53 crc kubenswrapper[4837]: I1001 07:57:53.823062 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:57:53 crc kubenswrapper[4837]: E1001 07:57:53.824292 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:58:07 crc kubenswrapper[4837]: I1001 07:58:07.816292 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:58:07 crc kubenswrapper[4837]: E1001 07:58:07.817818 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:58:18 crc kubenswrapper[4837]: I1001 07:58:18.815952 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:58:18 crc kubenswrapper[4837]: E1001 07:58:18.817159 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:58:29 crc kubenswrapper[4837]: I1001 07:58:29.816309 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:58:29 crc kubenswrapper[4837]: E1001 07:58:29.817132 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:58:42 crc kubenswrapper[4837]: I1001 07:58:42.816071 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:58:42 crc kubenswrapper[4837]: E1001 07:58:42.817202 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.120615 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.125993 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126048 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.126096 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126113 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.126135 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="extract-utilities" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126156 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="extract-utilities" Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.126194 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="extract-content" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126210 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="extract-content" Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.126250 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="extract-utilities" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126267 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="extract-utilities" Oct 01 07:58:45 crc kubenswrapper[4837]: E1001 07:58:45.126296 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="extract-content" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126312 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="extract-content" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126684 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3754cd-3b91-48d6-be0d-8af12e159cb1" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.126779 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f563afa7-3ca8-4c09-b0ac-030be2ce50f7" containerName="registry-server" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.129833 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.155119 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.313187 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.313248 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.313293 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p94d\" (UniqueName: \"kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.414830 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p94d\" (UniqueName: \"kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.415040 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.415094 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.415832 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.415883 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.454449 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p94d\" (UniqueName: \"kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d\") pod \"community-operators-gf9h6\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:45 crc kubenswrapper[4837]: I1001 07:58:45.460126 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:46 crc kubenswrapper[4837]: I1001 07:58:46.006767 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:46 crc kubenswrapper[4837]: I1001 07:58:46.664749 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerDied","Data":"97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d"} Oct 01 07:58:46 crc kubenswrapper[4837]: I1001 07:58:46.664916 4837 generic.go:334] "Generic (PLEG): container finished" podID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerID="97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d" exitCode=0 Oct 01 07:58:46 crc kubenswrapper[4837]: I1001 07:58:46.667267 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerStarted","Data":"61bd4b52515dc0d16a5eb95a7008aa37f4298a10d92b6ea2db3d911f0e412483"} Oct 01 07:58:46 crc kubenswrapper[4837]: I1001 07:58:46.668472 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 07:58:48 crc kubenswrapper[4837]: I1001 07:58:48.690799 4837 generic.go:334] "Generic (PLEG): container finished" podID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerID="2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a" exitCode=0 Oct 01 07:58:48 crc kubenswrapper[4837]: I1001 07:58:48.690915 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerDied","Data":"2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a"} Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.690231 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.692978 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.709565 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerStarted","Data":"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526"} Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.713097 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.763229 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gf9h6" podStartSLOduration=2.215526479 podStartE2EDuration="4.76321026s" podCreationTimestamp="2025-10-01 07:58:45 +0000 UTC" firstStartedPulling="2025-10-01 07:58:46.667996482 +0000 UTC m=+3183.509603977" lastFinishedPulling="2025-10-01 07:58:49.215680263 +0000 UTC m=+3186.057287758" observedRunningTime="2025-10-01 07:58:49.760246888 +0000 UTC m=+3186.601854433" watchObservedRunningTime="2025-10-01 07:58:49.76321026 +0000 UTC m=+3186.604817715" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.887052 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnw6v\" (UniqueName: \"kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.887508 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.887610 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.989537 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.989627 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.989726 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnw6v\" (UniqueName: \"kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.990322 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:49 crc kubenswrapper[4837]: I1001 07:58:49.990348 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.012059 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnw6v\" (UniqueName: \"kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v\") pod \"certified-operators-bkvkq\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.029289 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.274331 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.719042 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerID="bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847" exitCode=0 Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.720080 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerDied","Data":"bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847"} Oct 01 07:58:50 crc kubenswrapper[4837]: I1001 07:58:50.720140 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerStarted","Data":"136130f82cd07fa3246b7e4e253831817e6e7d4f39dfda106ed2f727417ec920"} Oct 01 07:58:52 crc kubenswrapper[4837]: I1001 07:58:52.753761 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerID="a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e" exitCode=0 Oct 01 07:58:52 crc kubenswrapper[4837]: I1001 07:58:52.754082 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerDied","Data":"a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e"} Oct 01 07:58:53 crc kubenswrapper[4837]: I1001 07:58:53.763880 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerStarted","Data":"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902"} Oct 01 07:58:53 crc kubenswrapper[4837]: I1001 07:58:53.789943 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bkvkq" podStartSLOduration=2.075653535 podStartE2EDuration="4.789910672s" podCreationTimestamp="2025-10-01 07:58:49 +0000 UTC" firstStartedPulling="2025-10-01 07:58:50.721341394 +0000 UTC m=+3187.562948889" lastFinishedPulling="2025-10-01 07:58:53.435598571 +0000 UTC m=+3190.277206026" observedRunningTime="2025-10-01 07:58:53.786001346 +0000 UTC m=+3190.627608861" watchObservedRunningTime="2025-10-01 07:58:53.789910672 +0000 UTC m=+3190.631518177" Oct 01 07:58:54 crc kubenswrapper[4837]: I1001 07:58:54.816575 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 07:58:55 crc kubenswrapper[4837]: I1001 07:58:55.461259 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:55 crc kubenswrapper[4837]: I1001 07:58:55.461879 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:55 crc kubenswrapper[4837]: I1001 07:58:55.537854 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:55 crc kubenswrapper[4837]: I1001 07:58:55.781082 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a"} Oct 01 07:58:55 crc kubenswrapper[4837]: I1001 07:58:55.835602 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:57 crc kubenswrapper[4837]: I1001 07:58:57.077167 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:58 crc kubenswrapper[4837]: I1001 07:58:58.822472 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gf9h6" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="registry-server" containerID="cri-o://9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526" gracePeriod=2 Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.308731 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.370327 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content\") pod \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.370438 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p94d\" (UniqueName: \"kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d\") pod \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.370472 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities\") pod \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\" (UID: \"13e134d3-4eeb-4f57-8d13-27ae5a89245c\") " Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.371394 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities" (OuterVolumeSpecName: "utilities") pod "13e134d3-4eeb-4f57-8d13-27ae5a89245c" (UID: "13e134d3-4eeb-4f57-8d13-27ae5a89245c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.378727 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d" (OuterVolumeSpecName: "kube-api-access-8p94d") pod "13e134d3-4eeb-4f57-8d13-27ae5a89245c" (UID: "13e134d3-4eeb-4f57-8d13-27ae5a89245c"). InnerVolumeSpecName "kube-api-access-8p94d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.436234 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13e134d3-4eeb-4f57-8d13-27ae5a89245c" (UID: "13e134d3-4eeb-4f57-8d13-27ae5a89245c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.471898 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.471929 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p94d\" (UniqueName: \"kubernetes.io/projected/13e134d3-4eeb-4f57-8d13-27ae5a89245c-kube-api-access-8p94d\") on node \"crc\" DevicePath \"\"" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.471939 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13e134d3-4eeb-4f57-8d13-27ae5a89245c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.838527 4837 generic.go:334] "Generic (PLEG): container finished" podID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerID="9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526" exitCode=0 Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.838616 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gf9h6" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.838646 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerDied","Data":"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526"} Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.839093 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gf9h6" event={"ID":"13e134d3-4eeb-4f57-8d13-27ae5a89245c","Type":"ContainerDied","Data":"61bd4b52515dc0d16a5eb95a7008aa37f4298a10d92b6ea2db3d911f0e412483"} Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.839145 4837 scope.go:117] "RemoveContainer" containerID="9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.870784 4837 scope.go:117] "RemoveContainer" containerID="2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.902375 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.907029 4837 scope.go:117] "RemoveContainer" containerID="97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.917891 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gf9h6"] Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.945356 4837 scope.go:117] "RemoveContainer" containerID="9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526" Oct 01 07:58:59 crc kubenswrapper[4837]: E1001 07:58:59.946026 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526\": container with ID starting with 9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526 not found: ID does not exist" containerID="9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.946077 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526"} err="failed to get container status \"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526\": rpc error: code = NotFound desc = could not find container \"9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526\": container with ID starting with 9e185900c5fea34acc549a36a2a1f237ede09633ce9f9cdca13a7230e117c526 not found: ID does not exist" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.946115 4837 scope.go:117] "RemoveContainer" containerID="2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a" Oct 01 07:58:59 crc kubenswrapper[4837]: E1001 07:58:59.946816 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a\": container with ID starting with 2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a not found: ID does not exist" containerID="2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.946883 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a"} err="failed to get container status \"2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a\": rpc error: code = NotFound desc = could not find container \"2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a\": container with ID starting with 2b257245dd4539e167fdb4026f023daea3327279970afd4b2dae3bfd4de13b0a not found: ID does not exist" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.946934 4837 scope.go:117] "RemoveContainer" containerID="97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d" Oct 01 07:58:59 crc kubenswrapper[4837]: E1001 07:58:59.947591 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d\": container with ID starting with 97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d not found: ID does not exist" containerID="97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d" Oct 01 07:58:59 crc kubenswrapper[4837]: I1001 07:58:59.947671 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d"} err="failed to get container status \"97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d\": rpc error: code = NotFound desc = could not find container \"97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d\": container with ID starting with 97e0a32b29dc29af596b882704ed52b3cd86ca86d3298f2dff7c7e654c34f05d not found: ID does not exist" Oct 01 07:59:00 crc kubenswrapper[4837]: I1001 07:59:00.030023 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:00 crc kubenswrapper[4837]: I1001 07:59:00.036116 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:00 crc kubenswrapper[4837]: I1001 07:59:00.111264 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:00 crc kubenswrapper[4837]: I1001 07:59:00.928260 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:01 crc kubenswrapper[4837]: I1001 07:59:01.838242 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" path="/var/lib/kubelet/pods/13e134d3-4eeb-4f57-8d13-27ae5a89245c/volumes" Oct 01 07:59:02 crc kubenswrapper[4837]: I1001 07:59:02.472120 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:59:03 crc kubenswrapper[4837]: I1001 07:59:03.881748 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bkvkq" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="registry-server" containerID="cri-o://c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902" gracePeriod=2 Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.469637 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.571423 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content\") pod \"3f227f5e-18fb-4496-9e08-9221b8050a82\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.571592 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnw6v\" (UniqueName: \"kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v\") pod \"3f227f5e-18fb-4496-9e08-9221b8050a82\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.571807 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities\") pod \"3f227f5e-18fb-4496-9e08-9221b8050a82\" (UID: \"3f227f5e-18fb-4496-9e08-9221b8050a82\") " Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.572832 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities" (OuterVolumeSpecName: "utilities") pod "3f227f5e-18fb-4496-9e08-9221b8050a82" (UID: "3f227f5e-18fb-4496-9e08-9221b8050a82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.578995 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v" (OuterVolumeSpecName: "kube-api-access-jnw6v") pod "3f227f5e-18fb-4496-9e08-9221b8050a82" (UID: "3f227f5e-18fb-4496-9e08-9221b8050a82"). InnerVolumeSpecName "kube-api-access-jnw6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.633684 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f227f5e-18fb-4496-9e08-9221b8050a82" (UID: "3f227f5e-18fb-4496-9e08-9221b8050a82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.674291 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnw6v\" (UniqueName: \"kubernetes.io/projected/3f227f5e-18fb-4496-9e08-9221b8050a82-kube-api-access-jnw6v\") on node \"crc\" DevicePath \"\"" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.674346 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.674366 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f227f5e-18fb-4496-9e08-9221b8050a82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.896462 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerID="c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902" exitCode=0 Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.896570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerDied","Data":"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902"} Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.896873 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkvkq" event={"ID":"3f227f5e-18fb-4496-9e08-9221b8050a82","Type":"ContainerDied","Data":"136130f82cd07fa3246b7e4e253831817e6e7d4f39dfda106ed2f727417ec920"} Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.896893 4837 scope.go:117] "RemoveContainer" containerID="c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.896645 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkvkq" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.923313 4837 scope.go:117] "RemoveContainer" containerID="a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.958931 4837 scope.go:117] "RemoveContainer" containerID="bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847" Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.960851 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:59:04 crc kubenswrapper[4837]: I1001 07:59:04.970313 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bkvkq"] Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.015684 4837 scope.go:117] "RemoveContainer" containerID="c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902" Oct 01 07:59:05 crc kubenswrapper[4837]: E1001 07:59:05.016180 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902\": container with ID starting with c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902 not found: ID does not exist" containerID="c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.016261 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902"} err="failed to get container status \"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902\": rpc error: code = NotFound desc = could not find container \"c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902\": container with ID starting with c120ce4c1074d1df0ab466a0bc7225c235b8f47fe9296e13acdb4108caa5f902 not found: ID does not exist" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.016322 4837 scope.go:117] "RemoveContainer" containerID="a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e" Oct 01 07:59:05 crc kubenswrapper[4837]: E1001 07:59:05.017050 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e\": container with ID starting with a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e not found: ID does not exist" containerID="a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.017102 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e"} err="failed to get container status \"a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e\": rpc error: code = NotFound desc = could not find container \"a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e\": container with ID starting with a3ac367f1e6a17e6da3a03f96ff238163b70b302541cf6d496c60c90b5ae1e8e not found: ID does not exist" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.017140 4837 scope.go:117] "RemoveContainer" containerID="bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847" Oct 01 07:59:05 crc kubenswrapper[4837]: E1001 07:59:05.017511 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847\": container with ID starting with bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847 not found: ID does not exist" containerID="bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.017575 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847"} err="failed to get container status \"bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847\": rpc error: code = NotFound desc = could not find container \"bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847\": container with ID starting with bc0151efa9a90f256cc2e6a958add3ae257f9738a6400705136f4c56902fe847 not found: ID does not exist" Oct 01 07:59:05 crc kubenswrapper[4837]: I1001 07:59:05.831537 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" path="/var/lib/kubelet/pods/3f227f5e-18fb-4496-9e08-9221b8050a82/volumes" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.211526 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d"] Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213042 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="extract-content" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213066 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="extract-content" Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213105 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213117 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213146 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="extract-content" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213166 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="extract-content" Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213202 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213213 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213228 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="extract-utilities" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213246 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="extract-utilities" Oct 01 08:00:00 crc kubenswrapper[4837]: E1001 08:00:00.213264 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="extract-utilities" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213274 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="extract-utilities" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213820 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e134d3-4eeb-4f57-8d13-27ae5a89245c" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.213851 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f227f5e-18fb-4496-9e08-9221b8050a82" containerName="registry-server" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.214930 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.219527 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.219573 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.231258 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d"] Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.381603 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.381770 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.381870 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdw4z\" (UniqueName: \"kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.485439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdw4z\" (UniqueName: \"kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.486191 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.486531 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.488947 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.496549 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.515252 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdw4z\" (UniqueName: \"kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z\") pod \"collect-profiles-29321760-x524d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.551826 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:00 crc kubenswrapper[4837]: I1001 08:00:00.817711 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d"] Oct 01 08:00:00 crc kubenswrapper[4837]: W1001 08:00:00.826896 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdc43b23_0756_47b2_94e8_267b8746967d.slice/crio-19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8 WatchSource:0}: Error finding container 19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8: Status 404 returned error can't find the container with id 19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8 Oct 01 08:00:01 crc kubenswrapper[4837]: I1001 08:00:01.438929 4837 generic.go:334] "Generic (PLEG): container finished" podID="fdc43b23-0756-47b2-94e8-267b8746967d" containerID="807e99129e0fd1f1d04280f79b08d2b4793b3c4fc4db205f6f81799243bfd259" exitCode=0 Oct 01 08:00:01 crc kubenswrapper[4837]: I1001 08:00:01.438976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" event={"ID":"fdc43b23-0756-47b2-94e8-267b8746967d","Type":"ContainerDied","Data":"807e99129e0fd1f1d04280f79b08d2b4793b3c4fc4db205f6f81799243bfd259"} Oct 01 08:00:01 crc kubenswrapper[4837]: I1001 08:00:01.439009 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" event={"ID":"fdc43b23-0756-47b2-94e8-267b8746967d","Type":"ContainerStarted","Data":"19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8"} Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.784931 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.928822 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume\") pod \"fdc43b23-0756-47b2-94e8-267b8746967d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.928925 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdw4z\" (UniqueName: \"kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z\") pod \"fdc43b23-0756-47b2-94e8-267b8746967d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.928991 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume\") pod \"fdc43b23-0756-47b2-94e8-267b8746967d\" (UID: \"fdc43b23-0756-47b2-94e8-267b8746967d\") " Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.929913 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume" (OuterVolumeSpecName: "config-volume") pod "fdc43b23-0756-47b2-94e8-267b8746967d" (UID: "fdc43b23-0756-47b2-94e8-267b8746967d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.930538 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdc43b23-0756-47b2-94e8-267b8746967d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.934656 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z" (OuterVolumeSpecName: "kube-api-access-cdw4z") pod "fdc43b23-0756-47b2-94e8-267b8746967d" (UID: "fdc43b23-0756-47b2-94e8-267b8746967d"). InnerVolumeSpecName "kube-api-access-cdw4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:00:02 crc kubenswrapper[4837]: I1001 08:00:02.935211 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fdc43b23-0756-47b2-94e8-267b8746967d" (UID: "fdc43b23-0756-47b2-94e8-267b8746967d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.032485 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdw4z\" (UniqueName: \"kubernetes.io/projected/fdc43b23-0756-47b2-94e8-267b8746967d-kube-api-access-cdw4z\") on node \"crc\" DevicePath \"\"" Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.032821 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdc43b23-0756-47b2-94e8-267b8746967d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.457497 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" event={"ID":"fdc43b23-0756-47b2-94e8-267b8746967d","Type":"ContainerDied","Data":"19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8"} Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.457789 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19f7a6d61ec56b681a3b1cb04c052ea2f8327ddf2e35fb0f92066bdced7f36b8" Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.457611 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d" Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.887072 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm"] Oct 01 08:00:03 crc kubenswrapper[4837]: I1001 08:00:03.892995 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321715-pmfnm"] Oct 01 08:00:05 crc kubenswrapper[4837]: I1001 08:00:05.837179 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d355c4fd-132f-45da-a9cd-94814064f59a" path="/var/lib/kubelet/pods/d355c4fd-132f-45da-a9cd-94814064f59a/volumes" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.148746 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:00:46 crc kubenswrapper[4837]: E1001 08:00:46.149967 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc43b23-0756-47b2-94e8-267b8746967d" containerName="collect-profiles" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.149989 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc43b23-0756-47b2-94e8-267b8746967d" containerName="collect-profiles" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.150242 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdc43b23-0756-47b2-94e8-267b8746967d" containerName="collect-profiles" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.152182 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.158279 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.180822 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.181094 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.181259 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvbxv\" (UniqueName: \"kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.283007 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.283484 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.283542 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvbxv\" (UniqueName: \"kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.283568 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.284189 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.321367 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvbxv\" (UniqueName: \"kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv\") pod \"redhat-marketplace-7smlq\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.495518 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:46 crc kubenswrapper[4837]: I1001 08:00:46.944393 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:00:47 crc kubenswrapper[4837]: I1001 08:00:47.891049 4837 generic.go:334] "Generic (PLEG): container finished" podID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerID="d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47" exitCode=0 Oct 01 08:00:47 crc kubenswrapper[4837]: I1001 08:00:47.891523 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerDied","Data":"d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47"} Oct 01 08:00:47 crc kubenswrapper[4837]: I1001 08:00:47.891578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerStarted","Data":"19a65295906b95990b9738f15fcf931db3dc5c23ec8927253254106ae4feadfe"} Oct 01 08:00:49 crc kubenswrapper[4837]: I1001 08:00:49.927244 4837 generic.go:334] "Generic (PLEG): container finished" podID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerID="c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713" exitCode=0 Oct 01 08:00:49 crc kubenswrapper[4837]: I1001 08:00:49.927560 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerDied","Data":"c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713"} Oct 01 08:00:50 crc kubenswrapper[4837]: I1001 08:00:50.948067 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerStarted","Data":"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a"} Oct 01 08:00:50 crc kubenswrapper[4837]: I1001 08:00:50.973407 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7smlq" podStartSLOduration=2.434292271 podStartE2EDuration="4.973383917s" podCreationTimestamp="2025-10-01 08:00:46 +0000 UTC" firstStartedPulling="2025-10-01 08:00:47.894347313 +0000 UTC m=+3304.735954808" lastFinishedPulling="2025-10-01 08:00:50.433438989 +0000 UTC m=+3307.275046454" observedRunningTime="2025-10-01 08:00:50.969092654 +0000 UTC m=+3307.810700149" watchObservedRunningTime="2025-10-01 08:00:50.973383917 +0000 UTC m=+3307.814991382" Oct 01 08:00:56 crc kubenswrapper[4837]: I1001 08:00:56.496775 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:56 crc kubenswrapper[4837]: I1001 08:00:56.497520 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:56 crc kubenswrapper[4837]: I1001 08:00:56.560519 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:57 crc kubenswrapper[4837]: I1001 08:00:57.100512 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:57 crc kubenswrapper[4837]: I1001 08:00:57.172729 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.041106 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7smlq" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="registry-server" containerID="cri-o://3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a" gracePeriod=2 Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.602117 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.799353 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content\") pod \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.799781 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities\") pod \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.799839 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvbxv\" (UniqueName: \"kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv\") pod \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\" (UID: \"6ba83d68-3752-4b2d-b7bc-5726d626e0cc\") " Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.801303 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities" (OuterVolumeSpecName: "utilities") pod "6ba83d68-3752-4b2d-b7bc-5726d626e0cc" (UID: "6ba83d68-3752-4b2d-b7bc-5726d626e0cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.808049 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv" (OuterVolumeSpecName: "kube-api-access-jvbxv") pod "6ba83d68-3752-4b2d-b7bc-5726d626e0cc" (UID: "6ba83d68-3752-4b2d-b7bc-5726d626e0cc"). InnerVolumeSpecName "kube-api-access-jvbxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.825416 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ba83d68-3752-4b2d-b7bc-5726d626e0cc" (UID: "6ba83d68-3752-4b2d-b7bc-5726d626e0cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.901309 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.901353 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:00:59 crc kubenswrapper[4837]: I1001 08:00:59.901383 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvbxv\" (UniqueName: \"kubernetes.io/projected/6ba83d68-3752-4b2d-b7bc-5726d626e0cc-kube-api-access-jvbxv\") on node \"crc\" DevicePath \"\"" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.048804 4837 generic.go:334] "Generic (PLEG): container finished" podID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerID="3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a" exitCode=0 Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.048847 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerDied","Data":"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a"} Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.048912 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7smlq" event={"ID":"6ba83d68-3752-4b2d-b7bc-5726d626e0cc","Type":"ContainerDied","Data":"19a65295906b95990b9738f15fcf931db3dc5c23ec8927253254106ae4feadfe"} Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.048915 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7smlq" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.048930 4837 scope.go:117] "RemoveContainer" containerID="3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.087852 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.088778 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7smlq"] Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.089663 4837 scope.go:117] "RemoveContainer" containerID="c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.118036 4837 scope.go:117] "RemoveContainer" containerID="d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.156088 4837 scope.go:117] "RemoveContainer" containerID="3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a" Oct 01 08:01:00 crc kubenswrapper[4837]: E1001 08:01:00.156535 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a\": container with ID starting with 3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a not found: ID does not exist" containerID="3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.156566 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a"} err="failed to get container status \"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a\": rpc error: code = NotFound desc = could not find container \"3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a\": container with ID starting with 3584d50207f7445e5b317fe170a4588f9ea7ad4461f25029a8068708e8c3947a not found: ID does not exist" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.156586 4837 scope.go:117] "RemoveContainer" containerID="c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713" Oct 01 08:01:00 crc kubenswrapper[4837]: E1001 08:01:00.156977 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713\": container with ID starting with c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713 not found: ID does not exist" containerID="c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.157042 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713"} err="failed to get container status \"c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713\": rpc error: code = NotFound desc = could not find container \"c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713\": container with ID starting with c8c7cbff693e02c929c6ebd010a799feaa24e4a40780591acb0252156d0d6713 not found: ID does not exist" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.157072 4837 scope.go:117] "RemoveContainer" containerID="d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47" Oct 01 08:01:00 crc kubenswrapper[4837]: E1001 08:01:00.157412 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47\": container with ID starting with d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47 not found: ID does not exist" containerID="d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47" Oct 01 08:01:00 crc kubenswrapper[4837]: I1001 08:01:00.157433 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47"} err="failed to get container status \"d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47\": rpc error: code = NotFound desc = could not find container \"d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47\": container with ID starting with d05d8b9882a737f158969b6dd62d0e138e21c66dc8adcc0afeffbc6add7e1f47 not found: ID does not exist" Oct 01 08:01:01 crc kubenswrapper[4837]: I1001 08:01:01.832884 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" path="/var/lib/kubelet/pods/6ba83d68-3752-4b2d-b7bc-5726d626e0cc/volumes" Oct 01 08:01:03 crc kubenswrapper[4837]: I1001 08:01:03.802959 4837 scope.go:117] "RemoveContainer" containerID="ea9a586dd5ca02047896b806fc47eaec453fedcf10a74898cb12ba92523202ce" Oct 01 08:01:23 crc kubenswrapper[4837]: I1001 08:01:23.080225 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:01:23 crc kubenswrapper[4837]: I1001 08:01:23.080952 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.920607 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:28 crc kubenswrapper[4837]: E1001 08:01:28.928467 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="extract-content" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.929063 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="extract-content" Oct 01 08:01:28 crc kubenswrapper[4837]: E1001 08:01:28.929226 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="extract-utilities" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.929388 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="extract-utilities" Oct 01 08:01:28 crc kubenswrapper[4837]: E1001 08:01:28.929545 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="registry-server" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.929776 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="registry-server" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.930271 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba83d68-3752-4b2d-b7bc-5726d626e0cc" containerName="registry-server" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.932630 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:28 crc kubenswrapper[4837]: I1001 08:01:28.947213 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.082543 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.082645 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlpdl\" (UniqueName: \"kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.082859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.184497 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.184548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlpdl\" (UniqueName: \"kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.184605 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.185074 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.185087 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.210234 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlpdl\" (UniqueName: \"kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl\") pod \"redhat-operators-lxmzz\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.269861 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:29 crc kubenswrapper[4837]: I1001 08:01:29.735312 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:30 crc kubenswrapper[4837]: I1001 08:01:30.328609 4837 generic.go:334] "Generic (PLEG): container finished" podID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerID="8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871" exitCode=0 Oct 01 08:01:30 crc kubenswrapper[4837]: I1001 08:01:30.328752 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerDied","Data":"8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871"} Oct 01 08:01:30 crc kubenswrapper[4837]: I1001 08:01:30.329039 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerStarted","Data":"ec3e5fbb234d06cd687cada13f85799ba6f372dee5a1bd60967dd12af2141c05"} Oct 01 08:01:33 crc kubenswrapper[4837]: I1001 08:01:33.368826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerStarted","Data":"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d"} Oct 01 08:01:34 crc kubenswrapper[4837]: I1001 08:01:34.384139 4837 generic.go:334] "Generic (PLEG): container finished" podID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerID="4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d" exitCode=0 Oct 01 08:01:34 crc kubenswrapper[4837]: I1001 08:01:34.384201 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerDied","Data":"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d"} Oct 01 08:01:42 crc kubenswrapper[4837]: I1001 08:01:42.474573 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerStarted","Data":"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6"} Oct 01 08:01:42 crc kubenswrapper[4837]: I1001 08:01:42.531253 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lxmzz" podStartSLOduration=3.361547905 podStartE2EDuration="14.53122671s" podCreationTimestamp="2025-10-01 08:01:28 +0000 UTC" firstStartedPulling="2025-10-01 08:01:30.331165489 +0000 UTC m=+3347.172772984" lastFinishedPulling="2025-10-01 08:01:41.500844294 +0000 UTC m=+3358.342451789" observedRunningTime="2025-10-01 08:01:42.517772385 +0000 UTC m=+3359.359379930" watchObservedRunningTime="2025-10-01 08:01:42.53122671 +0000 UTC m=+3359.372834205" Oct 01 08:01:49 crc kubenswrapper[4837]: I1001 08:01:49.270444 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:49 crc kubenswrapper[4837]: I1001 08:01:49.271096 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:49 crc kubenswrapper[4837]: I1001 08:01:49.324104 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:49 crc kubenswrapper[4837]: I1001 08:01:49.613338 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:49 crc kubenswrapper[4837]: I1001 08:01:49.680636 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:51 crc kubenswrapper[4837]: I1001 08:01:51.555914 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lxmzz" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="registry-server" containerID="cri-o://246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6" gracePeriod=2 Oct 01 08:01:51 crc kubenswrapper[4837]: I1001 08:01:51.979995 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.084942 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlpdl\" (UniqueName: \"kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl\") pod \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.085092 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content\") pod \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.085172 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities\") pod \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\" (UID: \"9be4d100-6ab7-4de8-8a17-d4acfc4864c6\") " Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.086285 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities" (OuterVolumeSpecName: "utilities") pod "9be4d100-6ab7-4de8-8a17-d4acfc4864c6" (UID: "9be4d100-6ab7-4de8-8a17-d4acfc4864c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.095349 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl" (OuterVolumeSpecName: "kube-api-access-mlpdl") pod "9be4d100-6ab7-4de8-8a17-d4acfc4864c6" (UID: "9be4d100-6ab7-4de8-8a17-d4acfc4864c6"). InnerVolumeSpecName "kube-api-access-mlpdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.187206 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.187248 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlpdl\" (UniqueName: \"kubernetes.io/projected/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-kube-api-access-mlpdl\") on node \"crc\" DevicePath \"\"" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.194372 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9be4d100-6ab7-4de8-8a17-d4acfc4864c6" (UID: "9be4d100-6ab7-4de8-8a17-d4acfc4864c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.288895 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9be4d100-6ab7-4de8-8a17-d4acfc4864c6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.568888 4837 generic.go:334] "Generic (PLEG): container finished" podID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerID="246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6" exitCode=0 Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.569009 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxmzz" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.568951 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerDied","Data":"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6"} Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.569796 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxmzz" event={"ID":"9be4d100-6ab7-4de8-8a17-d4acfc4864c6","Type":"ContainerDied","Data":"ec3e5fbb234d06cd687cada13f85799ba6f372dee5a1bd60967dd12af2141c05"} Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.569856 4837 scope.go:117] "RemoveContainer" containerID="246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.615941 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.620838 4837 scope.go:117] "RemoveContainer" containerID="4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.629539 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lxmzz"] Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.647768 4837 scope.go:117] "RemoveContainer" containerID="8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.674474 4837 scope.go:117] "RemoveContainer" containerID="246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6" Oct 01 08:01:52 crc kubenswrapper[4837]: E1001 08:01:52.674961 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6\": container with ID starting with 246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6 not found: ID does not exist" containerID="246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.674999 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6"} err="failed to get container status \"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6\": rpc error: code = NotFound desc = could not find container \"246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6\": container with ID starting with 246127be421693574e686de67319e31f22818526fd2266d38f8649233e9640d6 not found: ID does not exist" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.675020 4837 scope.go:117] "RemoveContainer" containerID="4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d" Oct 01 08:01:52 crc kubenswrapper[4837]: E1001 08:01:52.675384 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d\": container with ID starting with 4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d not found: ID does not exist" containerID="4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.675433 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d"} err="failed to get container status \"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d\": rpc error: code = NotFound desc = could not find container \"4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d\": container with ID starting with 4139fae7de41fe1f7e2ffb4ddd1169d622f1910faf1243565c6bb64c3d8a246d not found: ID does not exist" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.675472 4837 scope.go:117] "RemoveContainer" containerID="8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871" Oct 01 08:01:52 crc kubenswrapper[4837]: E1001 08:01:52.675892 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871\": container with ID starting with 8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871 not found: ID does not exist" containerID="8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871" Oct 01 08:01:52 crc kubenswrapper[4837]: I1001 08:01:52.675921 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871"} err="failed to get container status \"8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871\": rpc error: code = NotFound desc = could not find container \"8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871\": container with ID starting with 8d24c8d7569be7b7341919d4eea61b092d4280f8a156d7065aec52adc2361871 not found: ID does not exist" Oct 01 08:01:53 crc kubenswrapper[4837]: I1001 08:01:53.079630 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:01:53 crc kubenswrapper[4837]: I1001 08:01:53.080074 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:01:53 crc kubenswrapper[4837]: I1001 08:01:53.830739 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" path="/var/lib/kubelet/pods/9be4d100-6ab7-4de8-8a17-d4acfc4864c6/volumes" Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.079646 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.080230 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.080319 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.081081 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.081164 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a" gracePeriod=600 Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.862955 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a" exitCode=0 Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.863025 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a"} Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.863413 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f"} Oct 01 08:02:23 crc kubenswrapper[4837]: I1001 08:02:23.863439 4837 scope.go:117] "RemoveContainer" containerID="d3a0355f6ee106f11d82065d5801f7babb51ec368ac7ec5114f319d200446326" Oct 01 08:04:23 crc kubenswrapper[4837]: I1001 08:04:23.079362 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:04:23 crc kubenswrapper[4837]: I1001 08:04:23.080050 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:04:53 crc kubenswrapper[4837]: I1001 08:04:53.079513 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:04:53 crc kubenswrapper[4837]: I1001 08:04:53.080289 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.080285 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.081071 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.081141 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.082173 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.082311 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" gracePeriod=600 Oct 01 08:05:23 crc kubenswrapper[4837]: E1001 08:05:23.216459 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.522861 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" exitCode=0 Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.522929 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f"} Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.523014 4837 scope.go:117] "RemoveContainer" containerID="690e6d4c3db15c38e586a49c0a78bf14ef6d6e657851b463f22a84653749b70a" Oct 01 08:05:23 crc kubenswrapper[4837]: I1001 08:05:23.523995 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:05:23 crc kubenswrapper[4837]: E1001 08:05:23.524468 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:05:34 crc kubenswrapper[4837]: I1001 08:05:34.816246 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:05:34 crc kubenswrapper[4837]: E1001 08:05:34.817146 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:05:46 crc kubenswrapper[4837]: I1001 08:05:46.816377 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:05:46 crc kubenswrapper[4837]: E1001 08:05:46.817413 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:05:57 crc kubenswrapper[4837]: I1001 08:05:57.816156 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:05:57 crc kubenswrapper[4837]: E1001 08:05:57.817271 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:06:10 crc kubenswrapper[4837]: I1001 08:06:10.816764 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:06:10 crc kubenswrapper[4837]: E1001 08:06:10.817608 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:06:25 crc kubenswrapper[4837]: I1001 08:06:25.816774 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:06:25 crc kubenswrapper[4837]: E1001 08:06:25.817612 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:06:36 crc kubenswrapper[4837]: I1001 08:06:36.817354 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:06:36 crc kubenswrapper[4837]: E1001 08:06:36.818441 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:06:47 crc kubenswrapper[4837]: I1001 08:06:47.816971 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:06:47 crc kubenswrapper[4837]: E1001 08:06:47.818419 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:07:00 crc kubenswrapper[4837]: I1001 08:07:00.815896 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:07:00 crc kubenswrapper[4837]: E1001 08:07:00.816944 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:07:15 crc kubenswrapper[4837]: I1001 08:07:15.823031 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:07:15 crc kubenswrapper[4837]: E1001 08:07:15.823858 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:07:30 crc kubenswrapper[4837]: I1001 08:07:30.816386 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:07:30 crc kubenswrapper[4837]: E1001 08:07:30.817843 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:07:43 crc kubenswrapper[4837]: I1001 08:07:43.822137 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:07:43 crc kubenswrapper[4837]: E1001 08:07:43.823039 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:07:56 crc kubenswrapper[4837]: I1001 08:07:56.816435 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:07:56 crc kubenswrapper[4837]: E1001 08:07:56.817136 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:08:07 crc kubenswrapper[4837]: I1001 08:08:07.817899 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:08:07 crc kubenswrapper[4837]: E1001 08:08:07.818811 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:08:20 crc kubenswrapper[4837]: I1001 08:08:20.816920 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:08:20 crc kubenswrapper[4837]: E1001 08:08:20.818142 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:08:33 crc kubenswrapper[4837]: I1001 08:08:33.823222 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:08:33 crc kubenswrapper[4837]: E1001 08:08:33.824001 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:08:44 crc kubenswrapper[4837]: I1001 08:08:44.816640 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:08:44 crc kubenswrapper[4837]: E1001 08:08:44.817961 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:08:57 crc kubenswrapper[4837]: I1001 08:08:57.816963 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:08:57 crc kubenswrapper[4837]: E1001 08:08:57.819539 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:09:11 crc kubenswrapper[4837]: I1001 08:09:11.816810 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:09:11 crc kubenswrapper[4837]: E1001 08:09:11.817668 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:09:22 crc kubenswrapper[4837]: I1001 08:09:22.816231 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:09:22 crc kubenswrapper[4837]: E1001 08:09:22.818339 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:09:36 crc kubenswrapper[4837]: I1001 08:09:36.816435 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:09:36 crc kubenswrapper[4837]: E1001 08:09:36.819502 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:09:49 crc kubenswrapper[4837]: I1001 08:09:49.817169 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:09:49 crc kubenswrapper[4837]: E1001 08:09:49.818252 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:10:04 crc kubenswrapper[4837]: I1001 08:10:04.817077 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:10:04 crc kubenswrapper[4837]: E1001 08:10:04.818466 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:10:16 crc kubenswrapper[4837]: I1001 08:10:16.815861 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:10:16 crc kubenswrapper[4837]: E1001 08:10:16.817130 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:10:28 crc kubenswrapper[4837]: I1001 08:10:28.815934 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:10:29 crc kubenswrapper[4837]: I1001 08:10:29.474470 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758"} Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.039607 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:39 crc kubenswrapper[4837]: E1001 08:10:39.041363 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="extract-content" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.041398 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="extract-content" Oct 01 08:10:39 crc kubenswrapper[4837]: E1001 08:10:39.041443 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="extract-utilities" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.041460 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="extract-utilities" Oct 01 08:10:39 crc kubenswrapper[4837]: E1001 08:10:39.041504 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="registry-server" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.041523 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="registry-server" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.041955 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9be4d100-6ab7-4de8-8a17-d4acfc4864c6" containerName="registry-server" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.045589 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.068723 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.127214 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bvzg\" (UniqueName: \"kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.127553 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.127601 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.228975 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.229046 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.229122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bvzg\" (UniqueName: \"kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.229554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.229619 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.255317 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bvzg\" (UniqueName: \"kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg\") pod \"certified-operators-bt8rj\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.378494 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:39 crc kubenswrapper[4837]: I1001 08:10:39.942559 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.436890 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.439463 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.454368 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.550241 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.550366 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.550417 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv24r\" (UniqueName: \"kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.589421 4837 generic.go:334] "Generic (PLEG): container finished" podID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerID="9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01" exitCode=0 Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.589472 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerDied","Data":"9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01"} Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.589501 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerStarted","Data":"435523e0492c8344b16d47e8ac754d9930530780b900f00d2bb3324d5f77cd5f"} Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.591461 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.652186 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.652278 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.652334 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv24r\" (UniqueName: \"kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.653136 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.653164 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.672382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv24r\" (UniqueName: \"kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r\") pod \"community-operators-d55q9\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:40 crc kubenswrapper[4837]: I1001 08:10:40.768370 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.312327 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:41 crc kubenswrapper[4837]: W1001 08:10:41.319284 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf0c16da_343a_4323_b048_918e3eb814a3.slice/crio-0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37 WatchSource:0}: Error finding container 0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37: Status 404 returned error can't find the container with id 0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37 Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.599013 4837 generic.go:334] "Generic (PLEG): container finished" podID="bf0c16da-343a-4323-b048-918e3eb814a3" containerID="242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574" exitCode=0 Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.599124 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerDied","Data":"242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574"} Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.599166 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerStarted","Data":"0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37"} Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.600982 4837 generic.go:334] "Generic (PLEG): container finished" podID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerID="91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495" exitCode=0 Oct 01 08:10:41 crc kubenswrapper[4837]: I1001 08:10:41.601004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerDied","Data":"91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495"} Oct 01 08:10:42 crc kubenswrapper[4837]: I1001 08:10:42.610602 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerStarted","Data":"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab"} Oct 01 08:10:42 crc kubenswrapper[4837]: I1001 08:10:42.613136 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerStarted","Data":"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a"} Oct 01 08:10:42 crc kubenswrapper[4837]: I1001 08:10:42.660612 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bt8rj" podStartSLOduration=2.213845607 podStartE2EDuration="3.660585526s" podCreationTimestamp="2025-10-01 08:10:39 +0000 UTC" firstStartedPulling="2025-10-01 08:10:40.591126857 +0000 UTC m=+3897.432734312" lastFinishedPulling="2025-10-01 08:10:42.037866766 +0000 UTC m=+3898.879474231" observedRunningTime="2025-10-01 08:10:42.651050972 +0000 UTC m=+3899.492658467" watchObservedRunningTime="2025-10-01 08:10:42.660585526 +0000 UTC m=+3899.502193011" Oct 01 08:10:43 crc kubenswrapper[4837]: I1001 08:10:43.622758 4837 generic.go:334] "Generic (PLEG): container finished" podID="bf0c16da-343a-4323-b048-918e3eb814a3" containerID="247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab" exitCode=0 Oct 01 08:10:43 crc kubenswrapper[4837]: I1001 08:10:43.622819 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerDied","Data":"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab"} Oct 01 08:10:44 crc kubenswrapper[4837]: I1001 08:10:44.637176 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerStarted","Data":"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335"} Oct 01 08:10:44 crc kubenswrapper[4837]: I1001 08:10:44.659925 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d55q9" podStartSLOduration=2.043195676 podStartE2EDuration="4.65991068s" podCreationTimestamp="2025-10-01 08:10:40 +0000 UTC" firstStartedPulling="2025-10-01 08:10:41.600552404 +0000 UTC m=+3898.442159859" lastFinishedPulling="2025-10-01 08:10:44.217267398 +0000 UTC m=+3901.058874863" observedRunningTime="2025-10-01 08:10:44.657319496 +0000 UTC m=+3901.498926971" watchObservedRunningTime="2025-10-01 08:10:44.65991068 +0000 UTC m=+3901.501518135" Oct 01 08:10:49 crc kubenswrapper[4837]: I1001 08:10:49.379149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:49 crc kubenswrapper[4837]: I1001 08:10:49.380875 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:49 crc kubenswrapper[4837]: I1001 08:10:49.438863 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:49 crc kubenswrapper[4837]: I1001 08:10:49.754201 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:50 crc kubenswrapper[4837]: I1001 08:10:50.768911 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:50 crc kubenswrapper[4837]: I1001 08:10:50.769006 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:50 crc kubenswrapper[4837]: I1001 08:10:50.848938 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:51 crc kubenswrapper[4837]: I1001 08:10:51.425051 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:51 crc kubenswrapper[4837]: I1001 08:10:51.688848 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bt8rj" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="registry-server" containerID="cri-o://ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a" gracePeriod=2 Oct 01 08:10:51 crc kubenswrapper[4837]: I1001 08:10:51.753106 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.203350 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.346175 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities\") pod \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.346299 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content\") pod \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.346385 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bvzg\" (UniqueName: \"kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg\") pod \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\" (UID: \"d219423b-c8eb-4e2a-85b7-110e1cc8ae45\") " Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.346965 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities" (OuterVolumeSpecName: "utilities") pod "d219423b-c8eb-4e2a-85b7-110e1cc8ae45" (UID: "d219423b-c8eb-4e2a-85b7-110e1cc8ae45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.353564 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg" (OuterVolumeSpecName: "kube-api-access-2bvzg") pod "d219423b-c8eb-4e2a-85b7-110e1cc8ae45" (UID: "d219423b-c8eb-4e2a-85b7-110e1cc8ae45"). InnerVolumeSpecName "kube-api-access-2bvzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.448580 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.448621 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bvzg\" (UniqueName: \"kubernetes.io/projected/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-kube-api-access-2bvzg\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.700791 4837 generic.go:334] "Generic (PLEG): container finished" podID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerID="ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a" exitCode=0 Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.701010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerDied","Data":"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a"} Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.701563 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bt8rj" event={"ID":"d219423b-c8eb-4e2a-85b7-110e1cc8ae45","Type":"ContainerDied","Data":"435523e0492c8344b16d47e8ac754d9930530780b900f00d2bb3324d5f77cd5f"} Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.701590 4837 scope.go:117] "RemoveContainer" containerID="ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.701101 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bt8rj" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.720965 4837 scope.go:117] "RemoveContainer" containerID="91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.740815 4837 scope.go:117] "RemoveContainer" containerID="9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.770452 4837 scope.go:117] "RemoveContainer" containerID="ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a" Oct 01 08:10:52 crc kubenswrapper[4837]: E1001 08:10:52.771045 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a\": container with ID starting with ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a not found: ID does not exist" containerID="ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.771092 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a"} err="failed to get container status \"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a\": rpc error: code = NotFound desc = could not find container \"ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a\": container with ID starting with ce1cdba705c131d91c042dd17b1d4e5e220996a600496e09c80026f1d33e892a not found: ID does not exist" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.771124 4837 scope.go:117] "RemoveContainer" containerID="91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495" Oct 01 08:10:52 crc kubenswrapper[4837]: E1001 08:10:52.771582 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495\": container with ID starting with 91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495 not found: ID does not exist" containerID="91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.771645 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495"} err="failed to get container status \"91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495\": rpc error: code = NotFound desc = could not find container \"91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495\": container with ID starting with 91496d801b810da9c0264d7701d720e3c418b153b7e293708c2223051e1f1495 not found: ID does not exist" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.771732 4837 scope.go:117] "RemoveContainer" containerID="9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01" Oct 01 08:10:52 crc kubenswrapper[4837]: E1001 08:10:52.772193 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01\": container with ID starting with 9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01 not found: ID does not exist" containerID="9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.772235 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01"} err="failed to get container status \"9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01\": rpc error: code = NotFound desc = could not find container \"9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01\": container with ID starting with 9d0e96c714980bd56edd57063048a23555712c8a00ae5bf86c01f3b3c169ec01 not found: ID does not exist" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.837348 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d219423b-c8eb-4e2a-85b7-110e1cc8ae45" (UID: "d219423b-c8eb-4e2a-85b7-110e1cc8ae45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:10:52 crc kubenswrapper[4837]: I1001 08:10:52.858950 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d219423b-c8eb-4e2a-85b7-110e1cc8ae45-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:53 crc kubenswrapper[4837]: I1001 08:10:53.042626 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:53 crc kubenswrapper[4837]: I1001 08:10:53.042681 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bt8rj"] Oct 01 08:10:53 crc kubenswrapper[4837]: I1001 08:10:53.220216 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:53 crc kubenswrapper[4837]: I1001 08:10:53.712901 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d55q9" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="registry-server" containerID="cri-o://11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335" gracePeriod=2 Oct 01 08:10:53 crc kubenswrapper[4837]: I1001 08:10:53.823284 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" path="/var/lib/kubelet/pods/d219423b-c8eb-4e2a-85b7-110e1cc8ae45/volumes" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.101197 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.180386 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities\") pod \"bf0c16da-343a-4323-b048-918e3eb814a3\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.180947 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content\") pod \"bf0c16da-343a-4323-b048-918e3eb814a3\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.180969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv24r\" (UniqueName: \"kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r\") pod \"bf0c16da-343a-4323-b048-918e3eb814a3\" (UID: \"bf0c16da-343a-4323-b048-918e3eb814a3\") " Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.181349 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities" (OuterVolumeSpecName: "utilities") pod "bf0c16da-343a-4323-b048-918e3eb814a3" (UID: "bf0c16da-343a-4323-b048-918e3eb814a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.187983 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r" (OuterVolumeSpecName: "kube-api-access-qv24r") pod "bf0c16da-343a-4323-b048-918e3eb814a3" (UID: "bf0c16da-343a-4323-b048-918e3eb814a3"). InnerVolumeSpecName "kube-api-access-qv24r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.229645 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf0c16da-343a-4323-b048-918e3eb814a3" (UID: "bf0c16da-343a-4323-b048-918e3eb814a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.282083 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.282124 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv24r\" (UniqueName: \"kubernetes.io/projected/bf0c16da-343a-4323-b048-918e3eb814a3-kube-api-access-qv24r\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.282138 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf0c16da-343a-4323-b048-918e3eb814a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.729991 4837 generic.go:334] "Generic (PLEG): container finished" podID="bf0c16da-343a-4323-b048-918e3eb814a3" containerID="11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335" exitCode=0 Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.730096 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d55q9" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.730136 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerDied","Data":"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335"} Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.730919 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d55q9" event={"ID":"bf0c16da-343a-4323-b048-918e3eb814a3","Type":"ContainerDied","Data":"0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37"} Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.730957 4837 scope.go:117] "RemoveContainer" containerID="11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.761974 4837 scope.go:117] "RemoveContainer" containerID="247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.805003 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.813433 4837 scope.go:117] "RemoveContainer" containerID="242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.818220 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d55q9"] Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.842565 4837 scope.go:117] "RemoveContainer" containerID="11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335" Oct 01 08:10:54 crc kubenswrapper[4837]: E1001 08:10:54.843043 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335\": container with ID starting with 11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335 not found: ID does not exist" containerID="11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.843091 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335"} err="failed to get container status \"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335\": rpc error: code = NotFound desc = could not find container \"11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335\": container with ID starting with 11e11a729fd535a86be8ad1604abba6619c62e032bd4994d3f41567e3672c335 not found: ID does not exist" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.843124 4837 scope.go:117] "RemoveContainer" containerID="247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab" Oct 01 08:10:54 crc kubenswrapper[4837]: E1001 08:10:54.843431 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab\": container with ID starting with 247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab not found: ID does not exist" containerID="247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.843462 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab"} err="failed to get container status \"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab\": rpc error: code = NotFound desc = could not find container \"247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab\": container with ID starting with 247a131ca902f76779dddf01fc069922bc7ed5ab64bc05407fea10e787aa80ab not found: ID does not exist" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.843481 4837 scope.go:117] "RemoveContainer" containerID="242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574" Oct 01 08:10:54 crc kubenswrapper[4837]: E1001 08:10:54.843913 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574\": container with ID starting with 242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574 not found: ID does not exist" containerID="242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574" Oct 01 08:10:54 crc kubenswrapper[4837]: I1001 08:10:54.843967 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574"} err="failed to get container status \"242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574\": rpc error: code = NotFound desc = could not find container \"242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574\": container with ID starting with 242a8566b9f86629f094642fe7aa8130d5978ea9217eae731cdb83b94fa55574 not found: ID does not exist" Oct 01 08:10:54 crc kubenswrapper[4837]: E1001 08:10:54.915772 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf0c16da_343a_4323_b048_918e3eb814a3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf0c16da_343a_4323_b048_918e3eb814a3.slice/crio-0ae5923b3d29eec5b0cbfe66f17738821114a48b11c71e506a517a525ee1dc37\": RecentStats: unable to find data in memory cache]" Oct 01 08:10:55 crc kubenswrapper[4837]: I1001 08:10:55.827872 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" path="/var/lib/kubelet/pods/bf0c16da-343a-4323-b048-918e3eb814a3/volumes" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.198319 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199094 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="extract-content" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199472 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="extract-content" Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199490 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="extract-utilities" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199499 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="extract-utilities" Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199519 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="extract-content" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199528 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="extract-content" Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199542 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199550 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199567 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199575 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: E1001 08:12:37.199588 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="extract-utilities" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199595 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="extract-utilities" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199836 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d219423b-c8eb-4e2a-85b7-110e1cc8ae45" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.199856 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0c16da-343a-4323-b048-918e3eb814a3" containerName="registry-server" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.200999 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.230422 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.390811 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9z74\" (UniqueName: \"kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.390869 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.390931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.492277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9z74\" (UniqueName: \"kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.492327 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.492365 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.492867 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.493383 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.516747 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9z74\" (UniqueName: \"kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74\") pod \"redhat-operators-hnfvd\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.525825 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:37 crc kubenswrapper[4837]: I1001 08:12:37.988236 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:38 crc kubenswrapper[4837]: I1001 08:12:38.802275 4837 generic.go:334] "Generic (PLEG): container finished" podID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerID="7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee" exitCode=0 Oct 01 08:12:38 crc kubenswrapper[4837]: I1001 08:12:38.802433 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerDied","Data":"7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee"} Oct 01 08:12:38 crc kubenswrapper[4837]: I1001 08:12:38.802596 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerStarted","Data":"a5a4c1af0238ea6c1d717da1aeaeaba4b4659996bda08b39e30b78e754e00b7f"} Oct 01 08:12:39 crc kubenswrapper[4837]: I1001 08:12:39.814762 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerStarted","Data":"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c"} Oct 01 08:12:40 crc kubenswrapper[4837]: I1001 08:12:40.828102 4837 generic.go:334] "Generic (PLEG): container finished" podID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerID="2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c" exitCode=0 Oct 01 08:12:40 crc kubenswrapper[4837]: I1001 08:12:40.828180 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerDied","Data":"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c"} Oct 01 08:12:41 crc kubenswrapper[4837]: I1001 08:12:41.841258 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerStarted","Data":"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456"} Oct 01 08:12:41 crc kubenswrapper[4837]: I1001 08:12:41.873113 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hnfvd" podStartSLOduration=2.273662012 podStartE2EDuration="4.873051272s" podCreationTimestamp="2025-10-01 08:12:37 +0000 UTC" firstStartedPulling="2025-10-01 08:12:38.804511578 +0000 UTC m=+4015.646119033" lastFinishedPulling="2025-10-01 08:12:41.403900808 +0000 UTC m=+4018.245508293" observedRunningTime="2025-10-01 08:12:41.865907815 +0000 UTC m=+4018.707515330" watchObservedRunningTime="2025-10-01 08:12:41.873051272 +0000 UTC m=+4018.714658757" Oct 01 08:12:47 crc kubenswrapper[4837]: I1001 08:12:47.525955 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:47 crc kubenswrapper[4837]: I1001 08:12:47.526350 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:47 crc kubenswrapper[4837]: I1001 08:12:47.599785 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:47 crc kubenswrapper[4837]: I1001 08:12:47.958277 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:48 crc kubenswrapper[4837]: I1001 08:12:48.023165 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:49 crc kubenswrapper[4837]: I1001 08:12:49.913210 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hnfvd" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="registry-server" containerID="cri-o://89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456" gracePeriod=2 Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.421086 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.593991 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities\") pod \"4f5f0dc8-025e-4b38-96a3-d02425b58245\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.594152 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content\") pod \"4f5f0dc8-025e-4b38-96a3-d02425b58245\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.594200 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9z74\" (UniqueName: \"kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74\") pod \"4f5f0dc8-025e-4b38-96a3-d02425b58245\" (UID: \"4f5f0dc8-025e-4b38-96a3-d02425b58245\") " Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.595173 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities" (OuterVolumeSpecName: "utilities") pod "4f5f0dc8-025e-4b38-96a3-d02425b58245" (UID: "4f5f0dc8-025e-4b38-96a3-d02425b58245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.613987 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74" (OuterVolumeSpecName: "kube-api-access-w9z74") pod "4f5f0dc8-025e-4b38-96a3-d02425b58245" (UID: "4f5f0dc8-025e-4b38-96a3-d02425b58245"). InnerVolumeSpecName "kube-api-access-w9z74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.696096 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9z74\" (UniqueName: \"kubernetes.io/projected/4f5f0dc8-025e-4b38-96a3-d02425b58245-kube-api-access-w9z74\") on node \"crc\" DevicePath \"\"" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.696158 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.704130 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f5f0dc8-025e-4b38-96a3-d02425b58245" (UID: "4f5f0dc8-025e-4b38-96a3-d02425b58245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.798083 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f5f0dc8-025e-4b38-96a3-d02425b58245-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.926842 4837 generic.go:334] "Generic (PLEG): container finished" podID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerID="89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456" exitCode=0 Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.926891 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerDied","Data":"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456"} Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.926921 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnfvd" event={"ID":"4f5f0dc8-025e-4b38-96a3-d02425b58245","Type":"ContainerDied","Data":"a5a4c1af0238ea6c1d717da1aeaeaba4b4659996bda08b39e30b78e754e00b7f"} Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.926935 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnfvd" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.926942 4837 scope.go:117] "RemoveContainer" containerID="89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.957333 4837 scope.go:117] "RemoveContainer" containerID="2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c" Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.991670 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:50 crc kubenswrapper[4837]: I1001 08:12:50.998752 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hnfvd"] Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.011148 4837 scope.go:117] "RemoveContainer" containerID="7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.038005 4837 scope.go:117] "RemoveContainer" containerID="89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456" Oct 01 08:12:51 crc kubenswrapper[4837]: E1001 08:12:51.038906 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456\": container with ID starting with 89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456 not found: ID does not exist" containerID="89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.039035 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456"} err="failed to get container status \"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456\": rpc error: code = NotFound desc = could not find container \"89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456\": container with ID starting with 89c18bf4dd5e86e70f792fbe8503da3bb27070311dbd34f5a6ede859ff3a1456 not found: ID does not exist" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.039100 4837 scope.go:117] "RemoveContainer" containerID="2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c" Oct 01 08:12:51 crc kubenswrapper[4837]: E1001 08:12:51.039775 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c\": container with ID starting with 2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c not found: ID does not exist" containerID="2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.039833 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c"} err="failed to get container status \"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c\": rpc error: code = NotFound desc = could not find container \"2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c\": container with ID starting with 2ce7d135c2ff5a860d4555df95384759d271f5f8914d04c8f0f060e78ea7d68c not found: ID does not exist" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.039865 4837 scope.go:117] "RemoveContainer" containerID="7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee" Oct 01 08:12:51 crc kubenswrapper[4837]: E1001 08:12:51.040671 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee\": container with ID starting with 7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee not found: ID does not exist" containerID="7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.040740 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee"} err="failed to get container status \"7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee\": rpc error: code = NotFound desc = could not find container \"7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee\": container with ID starting with 7669c7a7907404447327fbac93e28303d6194d5ddab931982117a3635aeaffee not found: ID does not exist" Oct 01 08:12:51 crc kubenswrapper[4837]: I1001 08:12:51.834103 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" path="/var/lib/kubelet/pods/4f5f0dc8-025e-4b38-96a3-d02425b58245/volumes" Oct 01 08:12:53 crc kubenswrapper[4837]: I1001 08:12:53.079769 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:12:53 crc kubenswrapper[4837]: I1001 08:12:53.080100 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:13:23 crc kubenswrapper[4837]: I1001 08:13:23.080262 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:13:23 crc kubenswrapper[4837]: I1001 08:13:23.082277 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.079937 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.080677 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.080775 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.081591 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.081683 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758" gracePeriod=600 Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.531125 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758" exitCode=0 Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.531211 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758"} Oct 01 08:13:53 crc kubenswrapper[4837]: I1001 08:13:53.531332 4837 scope.go:117] "RemoveContainer" containerID="ec06f03da7be9ae8da4f6f31dcffa34d22a51a61222bb5712b28b6f7b7829e5f" Oct 01 08:13:54 crc kubenswrapper[4837]: I1001 08:13:54.545799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de"} Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.831553 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:14:52 crc kubenswrapper[4837]: E1001 08:14:52.833396 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="registry-server" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.833416 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="registry-server" Oct 01 08:14:52 crc kubenswrapper[4837]: E1001 08:14:52.833573 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="extract-utilities" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.833617 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="extract-utilities" Oct 01 08:14:52 crc kubenswrapper[4837]: E1001 08:14:52.833669 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="extract-content" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.833681 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="extract-content" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.834196 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5f0dc8-025e-4b38-96a3-d02425b58245" containerName="registry-server" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.835997 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.853242 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.994165 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.994233 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzfl\" (UniqueName: \"kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:52 crc kubenswrapper[4837]: I1001 08:14:52.994291 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.095407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.095478 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzfl\" (UniqueName: \"kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.095526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.096124 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.096586 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.122942 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzfl\" (UniqueName: \"kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl\") pod \"redhat-marketplace-cjfdt\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.186247 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:14:53 crc kubenswrapper[4837]: W1001 08:14:53.677936 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41b6cefc_7ce7_416f_8c0d_36f5335444be.slice/crio-2ce1fa1ce7d4249724150eced153e443b58d9d4bc93b8c9991c0d62bf3cf5b4f WatchSource:0}: Error finding container 2ce1fa1ce7d4249724150eced153e443b58d9d4bc93b8c9991c0d62bf3cf5b4f: Status 404 returned error can't find the container with id 2ce1fa1ce7d4249724150eced153e443b58d9d4bc93b8c9991c0d62bf3cf5b4f Oct 01 08:14:53 crc kubenswrapper[4837]: I1001 08:14:53.678289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:14:54 crc kubenswrapper[4837]: I1001 08:14:54.117192 4837 generic.go:334] "Generic (PLEG): container finished" podID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerID="4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e" exitCode=0 Oct 01 08:14:54 crc kubenswrapper[4837]: I1001 08:14:54.117353 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerDied","Data":"4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e"} Oct 01 08:14:54 crc kubenswrapper[4837]: I1001 08:14:54.117611 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerStarted","Data":"2ce1fa1ce7d4249724150eced153e443b58d9d4bc93b8c9991c0d62bf3cf5b4f"} Oct 01 08:14:56 crc kubenswrapper[4837]: I1001 08:14:56.147344 4837 generic.go:334] "Generic (PLEG): container finished" podID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerID="bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e" exitCode=0 Oct 01 08:14:56 crc kubenswrapper[4837]: I1001 08:14:56.147429 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerDied","Data":"bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e"} Oct 01 08:14:57 crc kubenswrapper[4837]: I1001 08:14:57.159649 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerStarted","Data":"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539"} Oct 01 08:14:57 crc kubenswrapper[4837]: I1001 08:14:57.177173 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cjfdt" podStartSLOduration=2.712791887 podStartE2EDuration="5.177150585s" podCreationTimestamp="2025-10-01 08:14:52 +0000 UTC" firstStartedPulling="2025-10-01 08:14:54.119672704 +0000 UTC m=+4150.961280199" lastFinishedPulling="2025-10-01 08:14:56.584031432 +0000 UTC m=+4153.425638897" observedRunningTime="2025-10-01 08:14:57.1749547 +0000 UTC m=+4154.016562285" watchObservedRunningTime="2025-10-01 08:14:57.177150585 +0000 UTC m=+4154.018758060" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.169260 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c"] Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.171012 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.176381 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.186729 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.191116 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c"] Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.372652 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.372981 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4v9l\" (UniqueName: \"kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.373049 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.474774 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.474907 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.474997 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4v9l\" (UniqueName: \"kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.476803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.483036 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.502073 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4v9l\" (UniqueName: \"kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l\") pod \"collect-profiles-29321775-6z86c\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:00 crc kubenswrapper[4837]: I1001 08:15:00.511356 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:01 crc kubenswrapper[4837]: I1001 08:15:01.000138 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c"] Oct 01 08:15:01 crc kubenswrapper[4837]: W1001 08:15:01.008862 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod261058d3_6d4a_4612_a6dc_6b0882515208.slice/crio-5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f WatchSource:0}: Error finding container 5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f: Status 404 returned error can't find the container with id 5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f Oct 01 08:15:01 crc kubenswrapper[4837]: I1001 08:15:01.199118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" event={"ID":"261058d3-6d4a-4612-a6dc-6b0882515208","Type":"ContainerStarted","Data":"5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f"} Oct 01 08:15:02 crc kubenswrapper[4837]: I1001 08:15:02.212610 4837 generic.go:334] "Generic (PLEG): container finished" podID="261058d3-6d4a-4612-a6dc-6b0882515208" containerID="775abe43fd64391df42b55ef1cabcdc7f23f6e10ef7f845ed4d53d3b4a37f624" exitCode=0 Oct 01 08:15:02 crc kubenswrapper[4837]: I1001 08:15:02.212794 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" event={"ID":"261058d3-6d4a-4612-a6dc-6b0882515208","Type":"ContainerDied","Data":"775abe43fd64391df42b55ef1cabcdc7f23f6e10ef7f845ed4d53d3b4a37f624"} Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.187207 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.187852 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.284264 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.375330 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.566092 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.582909 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.741504 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume\") pod \"261058d3-6d4a-4612-a6dc-6b0882515208\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.741733 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume\") pod \"261058d3-6d4a-4612-a6dc-6b0882515208\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.741800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4v9l\" (UniqueName: \"kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l\") pod \"261058d3-6d4a-4612-a6dc-6b0882515208\" (UID: \"261058d3-6d4a-4612-a6dc-6b0882515208\") " Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.742940 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume" (OuterVolumeSpecName: "config-volume") pod "261058d3-6d4a-4612-a6dc-6b0882515208" (UID: "261058d3-6d4a-4612-a6dc-6b0882515208"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.747848 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l" (OuterVolumeSpecName: "kube-api-access-b4v9l") pod "261058d3-6d4a-4612-a6dc-6b0882515208" (UID: "261058d3-6d4a-4612-a6dc-6b0882515208"). InnerVolumeSpecName "kube-api-access-b4v9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.749333 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "261058d3-6d4a-4612-a6dc-6b0882515208" (UID: "261058d3-6d4a-4612-a6dc-6b0882515208"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.843661 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4v9l\" (UniqueName: \"kubernetes.io/projected/261058d3-6d4a-4612-a6dc-6b0882515208-kube-api-access-b4v9l\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.843978 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/261058d3-6d4a-4612-a6dc-6b0882515208-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:03 crc kubenswrapper[4837]: I1001 08:15:03.844063 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/261058d3-6d4a-4612-a6dc-6b0882515208-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:04 crc kubenswrapper[4837]: I1001 08:15:04.234583 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" event={"ID":"261058d3-6d4a-4612-a6dc-6b0882515208","Type":"ContainerDied","Data":"5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f"} Oct 01 08:15:04 crc kubenswrapper[4837]: I1001 08:15:04.234647 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ebc23e746b79728075e0723f9cf327d55c6fcd0dbc6a9b0201eaeddc155fd4f" Oct 01 08:15:04 crc kubenswrapper[4837]: I1001 08:15:04.234767 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c" Oct 01 08:15:04 crc kubenswrapper[4837]: I1001 08:15:04.680255 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj"] Oct 01 08:15:04 crc kubenswrapper[4837]: I1001 08:15:04.689102 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321730-ldphj"] Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.243510 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cjfdt" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="registry-server" containerID="cri-o://49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539" gracePeriod=2 Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.764758 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.839910 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="add6774c-ba3a-4da8-bc3b-ee0db048329b" path="/var/lib/kubelet/pods/add6774c-ba3a-4da8-bc3b-ee0db048329b/volumes" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.876996 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content\") pod \"41b6cefc-7ce7-416f-8c0d-36f5335444be\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.877054 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzzfl\" (UniqueName: \"kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl\") pod \"41b6cefc-7ce7-416f-8c0d-36f5335444be\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.877098 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities\") pod \"41b6cefc-7ce7-416f-8c0d-36f5335444be\" (UID: \"41b6cefc-7ce7-416f-8c0d-36f5335444be\") " Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.879367 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities" (OuterVolumeSpecName: "utilities") pod "41b6cefc-7ce7-416f-8c0d-36f5335444be" (UID: "41b6cefc-7ce7-416f-8c0d-36f5335444be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.885286 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl" (OuterVolumeSpecName: "kube-api-access-vzzfl") pod "41b6cefc-7ce7-416f-8c0d-36f5335444be" (UID: "41b6cefc-7ce7-416f-8c0d-36f5335444be"). InnerVolumeSpecName "kube-api-access-vzzfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.906439 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41b6cefc-7ce7-416f-8c0d-36f5335444be" (UID: "41b6cefc-7ce7-416f-8c0d-36f5335444be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.980570 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.980617 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzzfl\" (UniqueName: \"kubernetes.io/projected/41b6cefc-7ce7-416f-8c0d-36f5335444be-kube-api-access-vzzfl\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:05 crc kubenswrapper[4837]: I1001 08:15:05.980638 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b6cefc-7ce7-416f-8c0d-36f5335444be-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.250913 4837 generic.go:334] "Generic (PLEG): container finished" podID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerID="49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539" exitCode=0 Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.250960 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerDied","Data":"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539"} Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.250989 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cjfdt" event={"ID":"41b6cefc-7ce7-416f-8c0d-36f5335444be","Type":"ContainerDied","Data":"2ce1fa1ce7d4249724150eced153e443b58d9d4bc93b8c9991c0d62bf3cf5b4f"} Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.251011 4837 scope.go:117] "RemoveContainer" containerID="49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.251149 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cjfdt" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.284414 4837 scope.go:117] "RemoveContainer" containerID="bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.298119 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.308863 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cjfdt"] Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.315479 4837 scope.go:117] "RemoveContainer" containerID="4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.348080 4837 scope.go:117] "RemoveContainer" containerID="49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539" Oct 01 08:15:06 crc kubenswrapper[4837]: E1001 08:15:06.348929 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539\": container with ID starting with 49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539 not found: ID does not exist" containerID="49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.348995 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539"} err="failed to get container status \"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539\": rpc error: code = NotFound desc = could not find container \"49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539\": container with ID starting with 49ec776d55af0c9cd845dc6bcc78dd85310731274400161ccecd73a2a555c539 not found: ID does not exist" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.349041 4837 scope.go:117] "RemoveContainer" containerID="bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e" Oct 01 08:15:06 crc kubenswrapper[4837]: E1001 08:15:06.349408 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e\": container with ID starting with bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e not found: ID does not exist" containerID="bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.349449 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e"} err="failed to get container status \"bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e\": rpc error: code = NotFound desc = could not find container \"bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e\": container with ID starting with bb4aaaa358e8935904771fb2cfde9b43c30accb004c813c2354f3103163ae70e not found: ID does not exist" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.349479 4837 scope.go:117] "RemoveContainer" containerID="4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e" Oct 01 08:15:06 crc kubenswrapper[4837]: E1001 08:15:06.349818 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e\": container with ID starting with 4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e not found: ID does not exist" containerID="4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e" Oct 01 08:15:06 crc kubenswrapper[4837]: I1001 08:15:06.349873 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e"} err="failed to get container status \"4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e\": rpc error: code = NotFound desc = could not find container \"4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e\": container with ID starting with 4bc751721e2313e4b5ff1717da496d08bc28ae0262a51796d9a562e0c3e1aa2e not found: ID does not exist" Oct 01 08:15:07 crc kubenswrapper[4837]: I1001 08:15:07.834416 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" path="/var/lib/kubelet/pods/41b6cefc-7ce7-416f-8c0d-36f5335444be/volumes" Oct 01 08:15:53 crc kubenswrapper[4837]: I1001 08:15:53.079518 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:15:53 crc kubenswrapper[4837]: I1001 08:15:53.080031 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:16:04 crc kubenswrapper[4837]: I1001 08:16:04.267151 4837 scope.go:117] "RemoveContainer" containerID="f8473a10b72c7d997561e536f90d2dacf3e5c5011a16e107138e5123ae45e42d" Oct 01 08:16:23 crc kubenswrapper[4837]: I1001 08:16:23.079466 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:16:23 crc kubenswrapper[4837]: I1001 08:16:23.079996 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.079594 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.080279 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.080350 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.081318 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.081414 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" gracePeriod=600 Oct 01 08:16:53 crc kubenswrapper[4837]: E1001 08:16:53.215061 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.285084 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" exitCode=0 Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.285155 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de"} Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.285211 4837 scope.go:117] "RemoveContainer" containerID="b2efdabe150a3d22baf99ee8bfbb86733ac40880668beb0b416b502767aef758" Oct 01 08:16:53 crc kubenswrapper[4837]: I1001 08:16:53.285983 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:16:53 crc kubenswrapper[4837]: E1001 08:16:53.286451 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:17:03 crc kubenswrapper[4837]: I1001 08:17:03.824217 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:17:03 crc kubenswrapper[4837]: E1001 08:17:03.825455 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:17:16 crc kubenswrapper[4837]: I1001 08:17:16.815968 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:17:16 crc kubenswrapper[4837]: E1001 08:17:16.817094 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:17:29 crc kubenswrapper[4837]: I1001 08:17:29.823296 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:17:29 crc kubenswrapper[4837]: E1001 08:17:29.824557 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:17:44 crc kubenswrapper[4837]: I1001 08:17:44.816668 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:17:44 crc kubenswrapper[4837]: E1001 08:17:44.818056 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:17:58 crc kubenswrapper[4837]: I1001 08:17:58.817852 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:17:58 crc kubenswrapper[4837]: E1001 08:17:58.819498 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:18:11 crc kubenswrapper[4837]: I1001 08:18:11.816279 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:18:11 crc kubenswrapper[4837]: E1001 08:18:11.817136 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:18:26 crc kubenswrapper[4837]: I1001 08:18:26.816070 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:18:26 crc kubenswrapper[4837]: E1001 08:18:26.817141 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:18:40 crc kubenswrapper[4837]: I1001 08:18:40.816118 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:18:40 crc kubenswrapper[4837]: E1001 08:18:40.817115 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:18:55 crc kubenswrapper[4837]: I1001 08:18:55.816270 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:18:55 crc kubenswrapper[4837]: E1001 08:18:55.817397 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:19:09 crc kubenswrapper[4837]: I1001 08:19:09.816970 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:19:09 crc kubenswrapper[4837]: E1001 08:19:09.817943 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:19:24 crc kubenswrapper[4837]: I1001 08:19:24.816631 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:19:24 crc kubenswrapper[4837]: E1001 08:19:24.817652 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:19:39 crc kubenswrapper[4837]: I1001 08:19:39.816229 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:19:39 crc kubenswrapper[4837]: E1001 08:19:39.817344 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:19:51 crc kubenswrapper[4837]: I1001 08:19:51.817094 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:19:51 crc kubenswrapper[4837]: E1001 08:19:51.819023 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:20:04 crc kubenswrapper[4837]: I1001 08:20:04.815928 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:20:04 crc kubenswrapper[4837]: E1001 08:20:04.816779 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:20:18 crc kubenswrapper[4837]: I1001 08:20:18.816115 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:20:18 crc kubenswrapper[4837]: E1001 08:20:18.817095 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:20:30 crc kubenswrapper[4837]: I1001 08:20:30.816302 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:20:30 crc kubenswrapper[4837]: E1001 08:20:30.817356 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:20:42 crc kubenswrapper[4837]: I1001 08:20:42.816596 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:20:42 crc kubenswrapper[4837]: E1001 08:20:42.817350 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:20:56 crc kubenswrapper[4837]: I1001 08:20:56.816036 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:20:56 crc kubenswrapper[4837]: E1001 08:20:56.816958 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.517838 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:01 crc kubenswrapper[4837]: E1001 08:21:01.519177 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261058d3-6d4a-4612-a6dc-6b0882515208" containerName="collect-profiles" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519200 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="261058d3-6d4a-4612-a6dc-6b0882515208" containerName="collect-profiles" Oct 01 08:21:01 crc kubenswrapper[4837]: E1001 08:21:01.519221 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="extract-content" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519233 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="extract-content" Oct 01 08:21:01 crc kubenswrapper[4837]: E1001 08:21:01.519256 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="registry-server" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519268 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="registry-server" Oct 01 08:21:01 crc kubenswrapper[4837]: E1001 08:21:01.519307 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="extract-utilities" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519319 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="extract-utilities" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519548 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b6cefc-7ce7-416f-8c0d-36f5335444be" containerName="registry-server" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.519585 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="261058d3-6d4a-4612-a6dc-6b0882515208" containerName="collect-profiles" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.521405 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.539034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.708862 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.708944 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.708981 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2hjj\" (UniqueName: \"kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.809824 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.809888 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2hjj\" (UniqueName: \"kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.809990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.810747 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.811215 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.849507 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2hjj\" (UniqueName: \"kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj\") pod \"community-operators-6qhgv\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:01 crc kubenswrapper[4837]: I1001 08:21:01.866311 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:02 crc kubenswrapper[4837]: I1001 08:21:02.175947 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:02 crc kubenswrapper[4837]: I1001 08:21:02.704848 4837 generic.go:334] "Generic (PLEG): container finished" podID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerID="a5023daa5810c9902ebda2128cd4424ce660b2e655fd2ff283b1cc2f5a11ad39" exitCode=0 Oct 01 08:21:02 crc kubenswrapper[4837]: I1001 08:21:02.704987 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerDied","Data":"a5023daa5810c9902ebda2128cd4424ce660b2e655fd2ff283b1cc2f5a11ad39"} Oct 01 08:21:02 crc kubenswrapper[4837]: I1001 08:21:02.705027 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerStarted","Data":"2a6a10be79b606e86917d3b8124e60e5f503a6750da7a0defe25d97570ecd92d"} Oct 01 08:21:02 crc kubenswrapper[4837]: I1001 08:21:02.707754 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:21:03 crc kubenswrapper[4837]: I1001 08:21:03.717789 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerStarted","Data":"5a551c4444a818b947e98a5322ac9d82603898f13f33696d6422a85687ead771"} Oct 01 08:21:04 crc kubenswrapper[4837]: I1001 08:21:04.730631 4837 generic.go:334] "Generic (PLEG): container finished" podID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerID="5a551c4444a818b947e98a5322ac9d82603898f13f33696d6422a85687ead771" exitCode=0 Oct 01 08:21:04 crc kubenswrapper[4837]: I1001 08:21:04.730715 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerDied","Data":"5a551c4444a818b947e98a5322ac9d82603898f13f33696d6422a85687ead771"} Oct 01 08:21:05 crc kubenswrapper[4837]: I1001 08:21:05.746516 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerStarted","Data":"8c7e49e1c64102cb5bcd753fe31c781cd911b3a52dff4785856412275cc773ad"} Oct 01 08:21:05 crc kubenswrapper[4837]: I1001 08:21:05.777273 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6qhgv" podStartSLOduration=2.134691604 podStartE2EDuration="4.777249848s" podCreationTimestamp="2025-10-01 08:21:01 +0000 UTC" firstStartedPulling="2025-10-01 08:21:02.707259459 +0000 UTC m=+4519.548866954" lastFinishedPulling="2025-10-01 08:21:05.349817703 +0000 UTC m=+4522.191425198" observedRunningTime="2025-10-01 08:21:05.771581509 +0000 UTC m=+4522.613189004" watchObservedRunningTime="2025-10-01 08:21:05.777249848 +0000 UTC m=+4522.618857313" Oct 01 08:21:11 crc kubenswrapper[4837]: I1001 08:21:11.816444 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:21:11 crc kubenswrapper[4837]: E1001 08:21:11.817537 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:21:11 crc kubenswrapper[4837]: I1001 08:21:11.867446 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:11 crc kubenswrapper[4837]: I1001 08:21:11.868065 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:11 crc kubenswrapper[4837]: I1001 08:21:11.932332 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:12 crc kubenswrapper[4837]: I1001 08:21:12.893977 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:12 crc kubenswrapper[4837]: I1001 08:21:12.971862 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:14 crc kubenswrapper[4837]: I1001 08:21:14.832347 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6qhgv" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="registry-server" containerID="cri-o://8c7e49e1c64102cb5bcd753fe31c781cd911b3a52dff4785856412275cc773ad" gracePeriod=2 Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.844762 4837 generic.go:334] "Generic (PLEG): container finished" podID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerID="8c7e49e1c64102cb5bcd753fe31c781cd911b3a52dff4785856412275cc773ad" exitCode=0 Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.844862 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerDied","Data":"8c7e49e1c64102cb5bcd753fe31c781cd911b3a52dff4785856412275cc773ad"} Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.845341 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qhgv" event={"ID":"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63","Type":"ContainerDied","Data":"2a6a10be79b606e86917d3b8124e60e5f503a6750da7a0defe25d97570ecd92d"} Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.845360 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a6a10be79b606e86917d3b8124e60e5f503a6750da7a0defe25d97570ecd92d" Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.887966 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.931420 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities\") pod \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.931537 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2hjj\" (UniqueName: \"kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj\") pod \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.931571 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content\") pod \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\" (UID: \"fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63\") " Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.934349 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities" (OuterVolumeSpecName: "utilities") pod "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" (UID: "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:21:15 crc kubenswrapper[4837]: I1001 08:21:15.938653 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj" (OuterVolumeSpecName: "kube-api-access-h2hjj") pod "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" (UID: "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63"). InnerVolumeSpecName "kube-api-access-h2hjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.007419 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" (UID: "fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.033219 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.033270 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2hjj\" (UniqueName: \"kubernetes.io/projected/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-kube-api-access-h2hjj\") on node \"crc\" DevicePath \"\"" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.033284 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.853039 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qhgv" Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.896328 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:16 crc kubenswrapper[4837]: I1001 08:21:16.905275 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6qhgv"] Oct 01 08:21:17 crc kubenswrapper[4837]: I1001 08:21:17.828852 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" path="/var/lib/kubelet/pods/fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63/volumes" Oct 01 08:21:22 crc kubenswrapper[4837]: I1001 08:21:22.815742 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:21:22 crc kubenswrapper[4837]: E1001 08:21:22.816529 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:21:34 crc kubenswrapper[4837]: I1001 08:21:34.815902 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:21:34 crc kubenswrapper[4837]: E1001 08:21:34.817027 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:21:46 crc kubenswrapper[4837]: I1001 08:21:46.816012 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:21:46 crc kubenswrapper[4837]: E1001 08:21:46.816669 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.853300 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:21:54 crc kubenswrapper[4837]: E1001 08:21:54.854669 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="registry-server" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.855228 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="registry-server" Oct 01 08:21:54 crc kubenswrapper[4837]: E1001 08:21:54.855272 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="extract-utilities" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.855289 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="extract-utilities" Oct 01 08:21:54 crc kubenswrapper[4837]: E1001 08:21:54.855355 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="extract-content" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.855373 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="extract-content" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.855892 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3bf1d7-a7fc-4c9c-b045-35e1fd12bd63" containerName="registry-server" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.858515 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:54 crc kubenswrapper[4837]: I1001 08:21:54.866601 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.017089 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rfr\" (UniqueName: \"kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.017138 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.017200 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.118541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rfr\" (UniqueName: \"kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.118594 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.118634 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.119096 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.119272 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.142463 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rfr\" (UniqueName: \"kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr\") pod \"certified-operators-7dztn\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.186198 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:21:55 crc kubenswrapper[4837]: I1001 08:21:55.781760 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:21:56 crc kubenswrapper[4837]: I1001 08:21:56.236121 4837 generic.go:334] "Generic (PLEG): container finished" podID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerID="e0d6ce90b1a2e2ac2472bd388ea9c73aed23dffabb31de57f6be513b8d13b6fb" exitCode=0 Oct 01 08:21:56 crc kubenswrapper[4837]: I1001 08:21:56.236234 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerDied","Data":"e0d6ce90b1a2e2ac2472bd388ea9c73aed23dffabb31de57f6be513b8d13b6fb"} Oct 01 08:21:56 crc kubenswrapper[4837]: I1001 08:21:56.236326 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerStarted","Data":"bab09edebfd90cbe46a5bc7f4050249e898df2a44f430ee80aa27c18b293db2f"} Oct 01 08:21:58 crc kubenswrapper[4837]: I1001 08:21:58.254239 4837 generic.go:334] "Generic (PLEG): container finished" podID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerID="3e9dbd35b643fa3a025ab96f092db11022cab49b18cb12a23a26e97623cc2e28" exitCode=0 Oct 01 08:21:58 crc kubenswrapper[4837]: I1001 08:21:58.254349 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerDied","Data":"3e9dbd35b643fa3a025ab96f092db11022cab49b18cb12a23a26e97623cc2e28"} Oct 01 08:21:58 crc kubenswrapper[4837]: I1001 08:21:58.816901 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:21:59 crc kubenswrapper[4837]: I1001 08:21:59.267664 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52"} Oct 01 08:21:59 crc kubenswrapper[4837]: I1001 08:21:59.271994 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerStarted","Data":"f3ffee755ec321a93894072cdb05a465c74d10bf7250c00f675e84aded9546b4"} Oct 01 08:21:59 crc kubenswrapper[4837]: I1001 08:21:59.335737 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7dztn" podStartSLOduration=2.671944921 podStartE2EDuration="5.335712905s" podCreationTimestamp="2025-10-01 08:21:54 +0000 UTC" firstStartedPulling="2025-10-01 08:21:56.240045197 +0000 UTC m=+4573.081652692" lastFinishedPulling="2025-10-01 08:21:58.903813181 +0000 UTC m=+4575.745420676" observedRunningTime="2025-10-01 08:21:59.322115241 +0000 UTC m=+4576.163722696" watchObservedRunningTime="2025-10-01 08:21:59.335712905 +0000 UTC m=+4576.177320360" Oct 01 08:22:05 crc kubenswrapper[4837]: I1001 08:22:05.186437 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:05 crc kubenswrapper[4837]: I1001 08:22:05.187046 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:05 crc kubenswrapper[4837]: I1001 08:22:05.241404 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:05 crc kubenswrapper[4837]: I1001 08:22:05.383733 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:05 crc kubenswrapper[4837]: I1001 08:22:05.483642 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:22:07 crc kubenswrapper[4837]: I1001 08:22:07.339443 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7dztn" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="registry-server" containerID="cri-o://f3ffee755ec321a93894072cdb05a465c74d10bf7250c00f675e84aded9546b4" gracePeriod=2 Oct 01 08:22:08 crc kubenswrapper[4837]: I1001 08:22:08.350578 4837 generic.go:334] "Generic (PLEG): container finished" podID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerID="f3ffee755ec321a93894072cdb05a465c74d10bf7250c00f675e84aded9546b4" exitCode=0 Oct 01 08:22:08 crc kubenswrapper[4837]: I1001 08:22:08.350678 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerDied","Data":"f3ffee755ec321a93894072cdb05a465c74d10bf7250c00f675e84aded9546b4"} Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.056757 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.164864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6rfr\" (UniqueName: \"kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr\") pod \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.164945 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content\") pod \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.165008 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities\") pod \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\" (UID: \"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae\") " Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.166124 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities" (OuterVolumeSpecName: "utilities") pod "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" (UID: "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.172951 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr" (OuterVolumeSpecName: "kube-api-access-p6rfr") pod "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" (UID: "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae"). InnerVolumeSpecName "kube-api-access-p6rfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.267034 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6rfr\" (UniqueName: \"kubernetes.io/projected/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-kube-api-access-p6rfr\") on node \"crc\" DevicePath \"\"" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.267762 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.359836 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7dztn" event={"ID":"8a2ce6fd-188b-4e56-b57d-88cddb6f3eae","Type":"ContainerDied","Data":"bab09edebfd90cbe46a5bc7f4050249e898df2a44f430ee80aa27c18b293db2f"} Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.359897 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7dztn" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.359900 4837 scope.go:117] "RemoveContainer" containerID="f3ffee755ec321a93894072cdb05a465c74d10bf7250c00f675e84aded9546b4" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.398401 4837 scope.go:117] "RemoveContainer" containerID="3e9dbd35b643fa3a025ab96f092db11022cab49b18cb12a23a26e97623cc2e28" Oct 01 08:22:09 crc kubenswrapper[4837]: I1001 08:22:09.415999 4837 scope.go:117] "RemoveContainer" containerID="e0d6ce90b1a2e2ac2472bd388ea9c73aed23dffabb31de57f6be513b8d13b6fb" Oct 01 08:22:10 crc kubenswrapper[4837]: I1001 08:22:10.126026 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" (UID: "8a2ce6fd-188b-4e56-b57d-88cddb6f3eae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:22:10 crc kubenswrapper[4837]: I1001 08:22:10.181880 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:22:10 crc kubenswrapper[4837]: I1001 08:22:10.316973 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:22:10 crc kubenswrapper[4837]: I1001 08:22:10.331451 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7dztn"] Oct 01 08:22:11 crc kubenswrapper[4837]: I1001 08:22:11.831576 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" path="/var/lib/kubelet/pods/8a2ce6fd-188b-4e56-b57d-88cddb6f3eae/volumes" Oct 01 08:22:38 crc kubenswrapper[4837]: E1001 08:22:38.329334 4837 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.514s" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.907498 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:06 crc kubenswrapper[4837]: E1001 08:23:06.908378 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="extract-content" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.908396 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="extract-content" Oct 01 08:23:06 crc kubenswrapper[4837]: E1001 08:23:06.908423 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="registry-server" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.908431 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="registry-server" Oct 01 08:23:06 crc kubenswrapper[4837]: E1001 08:23:06.908449 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="extract-utilities" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.908456 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="extract-utilities" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.908651 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a2ce6fd-188b-4e56-b57d-88cddb6f3eae" containerName="registry-server" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.944373 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.964111 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmvdb\" (UniqueName: \"kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.964155 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.964216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:06 crc kubenswrapper[4837]: I1001 08:23:06.970340 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.066175 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmvdb\" (UniqueName: \"kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.066222 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.066260 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.066787 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.068506 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.097168 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmvdb\" (UniqueName: \"kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb\") pod \"redhat-operators-xbhzt\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.290799 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.734078 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:07 crc kubenswrapper[4837]: I1001 08:23:07.962340 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerStarted","Data":"9ce51af6c5f02a2c0fd4067f5e897d810b1b1ba94b4457e52df9ea29f066e66a"} Oct 01 08:23:08 crc kubenswrapper[4837]: I1001 08:23:08.972130 4837 generic.go:334] "Generic (PLEG): container finished" podID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerID="201844331da706136b721af1c91d716ee119b7ca26f9df1e226214738b48c9d6" exitCode=0 Oct 01 08:23:08 crc kubenswrapper[4837]: I1001 08:23:08.972184 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerDied","Data":"201844331da706136b721af1c91d716ee119b7ca26f9df1e226214738b48c9d6"} Oct 01 08:23:12 crc kubenswrapper[4837]: I1001 08:23:12.001955 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerStarted","Data":"e7dcf18d33e415a45cf982ba1d333152c9f94f043a17598b84ed7d1ac1cc5352"} Oct 01 08:23:13 crc kubenswrapper[4837]: I1001 08:23:13.015609 4837 generic.go:334] "Generic (PLEG): container finished" podID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerID="e7dcf18d33e415a45cf982ba1d333152c9f94f043a17598b84ed7d1ac1cc5352" exitCode=0 Oct 01 08:23:13 crc kubenswrapper[4837]: I1001 08:23:13.015674 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerDied","Data":"e7dcf18d33e415a45cf982ba1d333152c9f94f043a17598b84ed7d1ac1cc5352"} Oct 01 08:23:14 crc kubenswrapper[4837]: I1001 08:23:14.026411 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerStarted","Data":"5edba3f11a1a4df4e204f3c39ce51e50690cae534d63761960d9b58a753d44cf"} Oct 01 08:23:15 crc kubenswrapper[4837]: I1001 08:23:15.055355 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xbhzt" podStartSLOduration=4.258975395 podStartE2EDuration="9.05533022s" podCreationTimestamp="2025-10-01 08:23:06 +0000 UTC" firstStartedPulling="2025-10-01 08:23:08.976091525 +0000 UTC m=+4645.817698980" lastFinishedPulling="2025-10-01 08:23:13.77244634 +0000 UTC m=+4650.614053805" observedRunningTime="2025-10-01 08:23:15.052773977 +0000 UTC m=+4651.894381462" watchObservedRunningTime="2025-10-01 08:23:15.05533022 +0000 UTC m=+4651.896937705" Oct 01 08:23:17 crc kubenswrapper[4837]: I1001 08:23:17.292864 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:17 crc kubenswrapper[4837]: I1001 08:23:17.294630 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:18 crc kubenswrapper[4837]: I1001 08:23:18.342660 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xbhzt" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="registry-server" probeResult="failure" output=< Oct 01 08:23:18 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 08:23:18 crc kubenswrapper[4837]: > Oct 01 08:23:27 crc kubenswrapper[4837]: I1001 08:23:27.362252 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:27 crc kubenswrapper[4837]: I1001 08:23:27.445820 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:27 crc kubenswrapper[4837]: I1001 08:23:27.610847 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:29 crc kubenswrapper[4837]: I1001 08:23:29.164042 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xbhzt" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="registry-server" containerID="cri-o://5edba3f11a1a4df4e204f3c39ce51e50690cae534d63761960d9b58a753d44cf" gracePeriod=2 Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.175136 4837 generic.go:334] "Generic (PLEG): container finished" podID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerID="5edba3f11a1a4df4e204f3c39ce51e50690cae534d63761960d9b58a753d44cf" exitCode=0 Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.175651 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerDied","Data":"5edba3f11a1a4df4e204f3c39ce51e50690cae534d63761960d9b58a753d44cf"} Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.175675 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbhzt" event={"ID":"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a","Type":"ContainerDied","Data":"9ce51af6c5f02a2c0fd4067f5e897d810b1b1ba94b4457e52df9ea29f066e66a"} Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.175746 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce51af6c5f02a2c0fd4067f5e897d810b1b1ba94b4457e52df9ea29f066e66a" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.176007 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.294579 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmvdb\" (UniqueName: \"kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb\") pod \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.294724 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities\") pod \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.294790 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content\") pod \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\" (UID: \"5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a\") " Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.295726 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities" (OuterVolumeSpecName: "utilities") pod "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" (UID: "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.303616 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb" (OuterVolumeSpecName: "kube-api-access-qmvdb") pod "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" (UID: "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a"). InnerVolumeSpecName "kube-api-access-qmvdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.396240 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmvdb\" (UniqueName: \"kubernetes.io/projected/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-kube-api-access-qmvdb\") on node \"crc\" DevicePath \"\"" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.396272 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.405532 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" (UID: "5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:23:30 crc kubenswrapper[4837]: I1001 08:23:30.496945 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:23:31 crc kubenswrapper[4837]: I1001 08:23:31.186240 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbhzt" Oct 01 08:23:31 crc kubenswrapper[4837]: I1001 08:23:31.244868 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:31 crc kubenswrapper[4837]: I1001 08:23:31.259660 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xbhzt"] Oct 01 08:23:31 crc kubenswrapper[4837]: I1001 08:23:31.829090 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" path="/var/lib/kubelet/pods/5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a/volumes" Oct 01 08:24:23 crc kubenswrapper[4837]: I1001 08:24:23.079588 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:24:23 crc kubenswrapper[4837]: I1001 08:24:23.080428 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.080170 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.080853 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.106838 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:24:53 crc kubenswrapper[4837]: E1001 08:24:53.108149 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="registry-server" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.108182 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="registry-server" Oct 01 08:24:53 crc kubenswrapper[4837]: E1001 08:24:53.108233 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="extract-content" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.108246 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="extract-content" Oct 01 08:24:53 crc kubenswrapper[4837]: E1001 08:24:53.108262 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="extract-utilities" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.108273 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="extract-utilities" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.108491 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae0cd29-2e9e-4ec1-9b71-bb8fbd78bd3a" containerName="registry-server" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.109870 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.119574 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.135728 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vz5t\" (UniqueName: \"kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.135803 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.135824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.237565 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vz5t\" (UniqueName: \"kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.237872 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.237984 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.238541 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.238560 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.261502 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vz5t\" (UniqueName: \"kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t\") pod \"redhat-marketplace-tkpxb\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.437379 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.896992 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:24:53 crc kubenswrapper[4837]: I1001 08:24:53.940984 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerStarted","Data":"8703237c2b7a5930a1950a330940ed23dcfb069a6ba0cfb27c4fea2950634fe8"} Oct 01 08:24:54 crc kubenswrapper[4837]: I1001 08:24:54.951573 4837 generic.go:334] "Generic (PLEG): container finished" podID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerID="177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f" exitCode=0 Oct 01 08:24:54 crc kubenswrapper[4837]: I1001 08:24:54.951642 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerDied","Data":"177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f"} Oct 01 08:24:55 crc kubenswrapper[4837]: I1001 08:24:55.962343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerStarted","Data":"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d"} Oct 01 08:24:56 crc kubenswrapper[4837]: I1001 08:24:56.972154 4837 generic.go:334] "Generic (PLEG): container finished" podID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerID="2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d" exitCode=0 Oct 01 08:24:56 crc kubenswrapper[4837]: I1001 08:24:56.972209 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerDied","Data":"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d"} Oct 01 08:24:57 crc kubenswrapper[4837]: I1001 08:24:57.982603 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerStarted","Data":"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a"} Oct 01 08:24:58 crc kubenswrapper[4837]: I1001 08:24:58.012998 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tkpxb" podStartSLOduration=2.32904908 podStartE2EDuration="5.012968678s" podCreationTimestamp="2025-10-01 08:24:53 +0000 UTC" firstStartedPulling="2025-10-01 08:24:54.953392055 +0000 UTC m=+4751.794999510" lastFinishedPulling="2025-10-01 08:24:57.637311653 +0000 UTC m=+4754.478919108" observedRunningTime="2025-10-01 08:24:58.006272484 +0000 UTC m=+4754.847879929" watchObservedRunningTime="2025-10-01 08:24:58.012968678 +0000 UTC m=+4754.854576173" Oct 01 08:25:03 crc kubenswrapper[4837]: I1001 08:25:03.437819 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:03 crc kubenswrapper[4837]: I1001 08:25:03.438214 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:03 crc kubenswrapper[4837]: I1001 08:25:03.485842 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:04 crc kubenswrapper[4837]: I1001 08:25:04.100376 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:04 crc kubenswrapper[4837]: I1001 08:25:04.167460 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.057571 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tkpxb" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="registry-server" containerID="cri-o://51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a" gracePeriod=2 Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.544123 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.652444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities\") pod \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.652564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vz5t\" (UniqueName: \"kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t\") pod \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.653147 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities" (OuterVolumeSpecName: "utilities") pod "c97a49c6-4b87-4626-be7d-c27d7f4b59db" (UID: "c97a49c6-4b87-4626-be7d-c27d7f4b59db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.653918 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content\") pod \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\" (UID: \"c97a49c6-4b87-4626-be7d-c27d7f4b59db\") " Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.654615 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.662641 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t" (OuterVolumeSpecName: "kube-api-access-2vz5t") pod "c97a49c6-4b87-4626-be7d-c27d7f4b59db" (UID: "c97a49c6-4b87-4626-be7d-c27d7f4b59db"). InnerVolumeSpecName "kube-api-access-2vz5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.669661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c97a49c6-4b87-4626-be7d-c27d7f4b59db" (UID: "c97a49c6-4b87-4626-be7d-c27d7f4b59db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.755367 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c97a49c6-4b87-4626-be7d-c27d7f4b59db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:25:06 crc kubenswrapper[4837]: I1001 08:25:06.755405 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vz5t\" (UniqueName: \"kubernetes.io/projected/c97a49c6-4b87-4626-be7d-c27d7f4b59db-kube-api-access-2vz5t\") on node \"crc\" DevicePath \"\"" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.069991 4837 generic.go:334] "Generic (PLEG): container finished" podID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerID="51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a" exitCode=0 Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.070038 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerDied","Data":"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a"} Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.070015 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkpxb" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.070133 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkpxb" event={"ID":"c97a49c6-4b87-4626-be7d-c27d7f4b59db","Type":"ContainerDied","Data":"8703237c2b7a5930a1950a330940ed23dcfb069a6ba0cfb27c4fea2950634fe8"} Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.070185 4837 scope.go:117] "RemoveContainer" containerID="51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.102012 4837 scope.go:117] "RemoveContainer" containerID="2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.116617 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.123590 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkpxb"] Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.146788 4837 scope.go:117] "RemoveContainer" containerID="177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.176900 4837 scope.go:117] "RemoveContainer" containerID="51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a" Oct 01 08:25:07 crc kubenswrapper[4837]: E1001 08:25:07.177635 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a\": container with ID starting with 51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a not found: ID does not exist" containerID="51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.177736 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a"} err="failed to get container status \"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a\": rpc error: code = NotFound desc = could not find container \"51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a\": container with ID starting with 51a99547657140a4369f98a8e9e86b463b867fd15485ce7de299bc990ef9958a not found: ID does not exist" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.177784 4837 scope.go:117] "RemoveContainer" containerID="2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d" Oct 01 08:25:07 crc kubenswrapper[4837]: E1001 08:25:07.178356 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d\": container with ID starting with 2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d not found: ID does not exist" containerID="2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.178422 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d"} err="failed to get container status \"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d\": rpc error: code = NotFound desc = could not find container \"2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d\": container with ID starting with 2fceceac72706bca4ba709e74f13fb97383bcf9237330dfecd4dc30339a6ef5d not found: ID does not exist" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.178471 4837 scope.go:117] "RemoveContainer" containerID="177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f" Oct 01 08:25:07 crc kubenswrapper[4837]: E1001 08:25:07.179217 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f\": container with ID starting with 177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f not found: ID does not exist" containerID="177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.179284 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f"} err="failed to get container status \"177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f\": rpc error: code = NotFound desc = could not find container \"177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f\": container with ID starting with 177cc34b46d1cd8c00929b882274b285c02fa10bbd2afe30f1dcddd0ce215a7f not found: ID does not exist" Oct 01 08:25:07 crc kubenswrapper[4837]: I1001 08:25:07.835652 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" path="/var/lib/kubelet/pods/c97a49c6-4b87-4626-be7d-c27d7f4b59db/volumes" Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.080158 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.080865 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.080926 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.081775 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.081864 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52" gracePeriod=600 Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.214382 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52" exitCode=0 Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.214451 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52"} Oct 01 08:25:23 crc kubenswrapper[4837]: I1001 08:25:23.214532 4837 scope.go:117] "RemoveContainer" containerID="93dc63d2655a0d3854d442430cfd22b0536e2fe9e14816881456b53b1bd143de" Oct 01 08:25:24 crc kubenswrapper[4837]: I1001 08:25:24.225788 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63"} Oct 01 08:27:04 crc kubenswrapper[4837]: I1001 08:27:04.593593 4837 scope.go:117] "RemoveContainer" containerID="5a551c4444a818b947e98a5322ac9d82603898f13f33696d6422a85687ead771" Oct 01 08:27:04 crc kubenswrapper[4837]: I1001 08:27:04.625814 4837 scope.go:117] "RemoveContainer" containerID="a5023daa5810c9902ebda2128cd4424ce660b2e655fd2ff283b1cc2f5a11ad39" Oct 01 08:27:23 crc kubenswrapper[4837]: I1001 08:27:23.079355 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:27:23 crc kubenswrapper[4837]: I1001 08:27:23.079996 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:27:53 crc kubenswrapper[4837]: I1001 08:27:53.080545 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:27:53 crc kubenswrapper[4837]: I1001 08:27:53.081243 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:28:04 crc kubenswrapper[4837]: I1001 08:28:04.699519 4837 scope.go:117] "RemoveContainer" containerID="8c7e49e1c64102cb5bcd753fe31c781cd911b3a52dff4785856412275cc773ad" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.079670 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.080392 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.080452 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.081235 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.081322 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" gracePeriod=600 Oct 01 08:28:23 crc kubenswrapper[4837]: E1001 08:28:23.241463 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.871775 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" exitCode=0 Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.871856 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63"} Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.871924 4837 scope.go:117] "RemoveContainer" containerID="bd36b8fc3237ca185417b7ffdc0268a6da17e0361882cd0e2d2314e240fbae52" Oct 01 08:28:23 crc kubenswrapper[4837]: I1001 08:28:23.873037 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:28:23 crc kubenswrapper[4837]: E1001 08:28:23.874573 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:28:37 crc kubenswrapper[4837]: I1001 08:28:37.815549 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:28:37 crc kubenswrapper[4837]: E1001 08:28:37.816555 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:28:49 crc kubenswrapper[4837]: I1001 08:28:49.815757 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:28:49 crc kubenswrapper[4837]: E1001 08:28:49.816667 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:29:04 crc kubenswrapper[4837]: I1001 08:29:04.816493 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:29:04 crc kubenswrapper[4837]: E1001 08:29:04.817526 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:29:19 crc kubenswrapper[4837]: I1001 08:29:19.816369 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:29:19 crc kubenswrapper[4837]: E1001 08:29:19.817629 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:29:34 crc kubenswrapper[4837]: I1001 08:29:34.816400 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:29:34 crc kubenswrapper[4837]: E1001 08:29:34.817549 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:29:49 crc kubenswrapper[4837]: I1001 08:29:49.816343 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:29:49 crc kubenswrapper[4837]: E1001 08:29:49.817131 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.172923 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7"] Oct 01 08:30:00 crc kubenswrapper[4837]: E1001 08:30:00.174188 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="extract-content" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.174224 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="extract-content" Oct 01 08:30:00 crc kubenswrapper[4837]: E1001 08:30:00.174294 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="registry-server" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.174313 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="registry-server" Oct 01 08:30:00 crc kubenswrapper[4837]: E1001 08:30:00.174349 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="extract-utilities" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.174368 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="extract-utilities" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.174769 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97a49c6-4b87-4626-be7d-c27d7f4b59db" containerName="registry-server" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.175805 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.178637 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.178768 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.189272 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7"] Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.274671 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfksc\" (UniqueName: \"kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.274875 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.275075 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.376007 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfksc\" (UniqueName: \"kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.376062 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.376105 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.378193 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.383431 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.411912 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfksc\" (UniqueName: \"kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc\") pod \"collect-profiles-29321790-fpbs7\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.527493 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:00 crc kubenswrapper[4837]: I1001 08:30:00.828065 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7"] Oct 01 08:30:01 crc kubenswrapper[4837]: I1001 08:30:01.807149 4837 generic.go:334] "Generic (PLEG): container finished" podID="30331db9-fb15-4dc8-a777-45201a20608f" containerID="d7070aa49b6860632cba07b25ba3b2170cd02587dce8b98332874bcbfc39ddd6" exitCode=0 Oct 01 08:30:01 crc kubenswrapper[4837]: I1001 08:30:01.807596 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" event={"ID":"30331db9-fb15-4dc8-a777-45201a20608f","Type":"ContainerDied","Data":"d7070aa49b6860632cba07b25ba3b2170cd02587dce8b98332874bcbfc39ddd6"} Oct 01 08:30:01 crc kubenswrapper[4837]: I1001 08:30:01.807634 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" event={"ID":"30331db9-fb15-4dc8-a777-45201a20608f","Type":"ContainerStarted","Data":"da248de46a233adc53a13e6db6e17eca1558d5da834b7912aa3e131233c30102"} Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.211114 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.320836 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume\") pod \"30331db9-fb15-4dc8-a777-45201a20608f\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.322002 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume" (OuterVolumeSpecName: "config-volume") pod "30331db9-fb15-4dc8-a777-45201a20608f" (UID: "30331db9-fb15-4dc8-a777-45201a20608f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.322075 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfksc\" (UniqueName: \"kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc\") pod \"30331db9-fb15-4dc8-a777-45201a20608f\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.322900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume\") pod \"30331db9-fb15-4dc8-a777-45201a20608f\" (UID: \"30331db9-fb15-4dc8-a777-45201a20608f\") " Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.323252 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30331db9-fb15-4dc8-a777-45201a20608f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.328369 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "30331db9-fb15-4dc8-a777-45201a20608f" (UID: "30331db9-fb15-4dc8-a777-45201a20608f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.329532 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc" (OuterVolumeSpecName: "kube-api-access-qfksc") pod "30331db9-fb15-4dc8-a777-45201a20608f" (UID: "30331db9-fb15-4dc8-a777-45201a20608f"). InnerVolumeSpecName "kube-api-access-qfksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.425002 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfksc\" (UniqueName: \"kubernetes.io/projected/30331db9-fb15-4dc8-a777-45201a20608f-kube-api-access-qfksc\") on node \"crc\" DevicePath \"\"" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.425037 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30331db9-fb15-4dc8-a777-45201a20608f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.826357 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:30:03 crc kubenswrapper[4837]: E1001 08:30:03.828224 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.840808 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" event={"ID":"30331db9-fb15-4dc8-a777-45201a20608f","Type":"ContainerDied","Data":"da248de46a233adc53a13e6db6e17eca1558d5da834b7912aa3e131233c30102"} Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.841316 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da248de46a233adc53a13e6db6e17eca1558d5da834b7912aa3e131233c30102" Oct 01 08:30:03 crc kubenswrapper[4837]: I1001 08:30:03.840931 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7" Oct 01 08:30:04 crc kubenswrapper[4837]: I1001 08:30:04.318533 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7"] Oct 01 08:30:04 crc kubenswrapper[4837]: I1001 08:30:04.328644 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321745-9h6l7"] Oct 01 08:30:04 crc kubenswrapper[4837]: I1001 08:30:04.793912 4837 scope.go:117] "RemoveContainer" containerID="5edba3f11a1a4df4e204f3c39ce51e50690cae534d63761960d9b58a753d44cf" Oct 01 08:30:04 crc kubenswrapper[4837]: I1001 08:30:04.819757 4837 scope.go:117] "RemoveContainer" containerID="e7dcf18d33e415a45cf982ba1d333152c9f94f043a17598b84ed7d1ac1cc5352" Oct 01 08:30:04 crc kubenswrapper[4837]: I1001 08:30:04.848472 4837 scope.go:117] "RemoveContainer" containerID="201844331da706136b721af1c91d716ee119b7ca26f9df1e226214738b48c9d6" Oct 01 08:30:05 crc kubenswrapper[4837]: I1001 08:30:05.834473 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc6a77a-70f6-4100-9197-a5e0aa1c4d46" path="/var/lib/kubelet/pods/ffc6a77a-70f6-4100-9197-a5e0aa1c4d46/volumes" Oct 01 08:30:14 crc kubenswrapper[4837]: I1001 08:30:14.816014 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:30:14 crc kubenswrapper[4837]: E1001 08:30:14.816838 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:30:26 crc kubenswrapper[4837]: I1001 08:30:26.815990 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:30:26 crc kubenswrapper[4837]: E1001 08:30:26.817084 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:30:37 crc kubenswrapper[4837]: I1001 08:30:37.815794 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:30:37 crc kubenswrapper[4837]: E1001 08:30:37.816564 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:30:51 crc kubenswrapper[4837]: I1001 08:30:51.815850 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:30:51 crc kubenswrapper[4837]: E1001 08:30:51.816621 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:31:02 crc kubenswrapper[4837]: I1001 08:31:02.816173 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:31:02 crc kubenswrapper[4837]: E1001 08:31:02.817107 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:31:04 crc kubenswrapper[4837]: I1001 08:31:04.925880 4837 scope.go:117] "RemoveContainer" containerID="4f4286aa037cb6b9afe186444d61e6a9d5a3792738d6df3eea1f39d79cb705dc" Oct 01 08:31:14 crc kubenswrapper[4837]: I1001 08:31:14.815648 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:31:14 crc kubenswrapper[4837]: E1001 08:31:14.816725 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:31:25 crc kubenswrapper[4837]: I1001 08:31:25.816209 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:31:25 crc kubenswrapper[4837]: E1001 08:31:25.816958 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:31:36 crc kubenswrapper[4837]: I1001 08:31:36.966934 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:36 crc kubenswrapper[4837]: E1001 08:31:36.967910 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30331db9-fb15-4dc8-a777-45201a20608f" containerName="collect-profiles" Oct 01 08:31:36 crc kubenswrapper[4837]: I1001 08:31:36.967933 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="30331db9-fb15-4dc8-a777-45201a20608f" containerName="collect-profiles" Oct 01 08:31:36 crc kubenswrapper[4837]: I1001 08:31:36.968196 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="30331db9-fb15-4dc8-a777-45201a20608f" containerName="collect-profiles" Oct 01 08:31:36 crc kubenswrapper[4837]: I1001 08:31:36.969912 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:36 crc kubenswrapper[4837]: I1001 08:31:36.978823 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.006939 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.007282 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.007535 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9tjq\" (UniqueName: \"kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.108172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9tjq\" (UniqueName: \"kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.108255 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.108331 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.109026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.109026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.144076 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9tjq\" (UniqueName: \"kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq\") pod \"community-operators-swrvb\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.325926 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.621724 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.708877 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerStarted","Data":"91b2cd7fe9f2ce0c16b6dc482f5066ccb6808acd40d0c9dec67e27558a212876"} Oct 01 08:31:37 crc kubenswrapper[4837]: I1001 08:31:37.816194 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:31:37 crc kubenswrapper[4837]: E1001 08:31:37.816550 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:31:38 crc kubenswrapper[4837]: I1001 08:31:38.721260 4837 generic.go:334] "Generic (PLEG): container finished" podID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerID="c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb" exitCode=0 Oct 01 08:31:38 crc kubenswrapper[4837]: I1001 08:31:38.722533 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerDied","Data":"c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb"} Oct 01 08:31:38 crc kubenswrapper[4837]: I1001 08:31:38.724744 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:31:40 crc kubenswrapper[4837]: I1001 08:31:40.742907 4837 generic.go:334] "Generic (PLEG): container finished" podID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerID="a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3" exitCode=0 Oct 01 08:31:40 crc kubenswrapper[4837]: I1001 08:31:40.743022 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerDied","Data":"a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3"} Oct 01 08:31:42 crc kubenswrapper[4837]: I1001 08:31:42.765353 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerStarted","Data":"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d"} Oct 01 08:31:42 crc kubenswrapper[4837]: I1001 08:31:42.791878 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swrvb" podStartSLOduration=3.608415014 podStartE2EDuration="6.791859684s" podCreationTimestamp="2025-10-01 08:31:36 +0000 UTC" firstStartedPulling="2025-10-01 08:31:38.724508373 +0000 UTC m=+5155.566115828" lastFinishedPulling="2025-10-01 08:31:41.907953033 +0000 UTC m=+5158.749560498" observedRunningTime="2025-10-01 08:31:42.786193615 +0000 UTC m=+5159.627801070" watchObservedRunningTime="2025-10-01 08:31:42.791859684 +0000 UTC m=+5159.633467139" Oct 01 08:31:47 crc kubenswrapper[4837]: I1001 08:31:47.327376 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:47 crc kubenswrapper[4837]: I1001 08:31:47.327804 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:47 crc kubenswrapper[4837]: I1001 08:31:47.408177 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:47 crc kubenswrapper[4837]: I1001 08:31:47.876832 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:47 crc kubenswrapper[4837]: I1001 08:31:47.960098 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:49 crc kubenswrapper[4837]: I1001 08:31:49.836995 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swrvb" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="registry-server" containerID="cri-o://814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d" gracePeriod=2 Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.299682 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.421443 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content\") pod \"61656303-39a3-4b92-9f2c-e5243a63ca21\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.421567 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9tjq\" (UniqueName: \"kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq\") pod \"61656303-39a3-4b92-9f2c-e5243a63ca21\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.421646 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities\") pod \"61656303-39a3-4b92-9f2c-e5243a63ca21\" (UID: \"61656303-39a3-4b92-9f2c-e5243a63ca21\") " Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.422870 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities" (OuterVolumeSpecName: "utilities") pod "61656303-39a3-4b92-9f2c-e5243a63ca21" (UID: "61656303-39a3-4b92-9f2c-e5243a63ca21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.429205 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq" (OuterVolumeSpecName: "kube-api-access-l9tjq") pod "61656303-39a3-4b92-9f2c-e5243a63ca21" (UID: "61656303-39a3-4b92-9f2c-e5243a63ca21"). InnerVolumeSpecName "kube-api-access-l9tjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.478980 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61656303-39a3-4b92-9f2c-e5243a63ca21" (UID: "61656303-39a3-4b92-9f2c-e5243a63ca21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.523799 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.523853 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61656303-39a3-4b92-9f2c-e5243a63ca21-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.523872 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9tjq\" (UniqueName: \"kubernetes.io/projected/61656303-39a3-4b92-9f2c-e5243a63ca21-kube-api-access-l9tjq\") on node \"crc\" DevicePath \"\"" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.849369 4837 generic.go:334] "Generic (PLEG): container finished" podID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerID="814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d" exitCode=0 Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.849427 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerDied","Data":"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d"} Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.849466 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swrvb" event={"ID":"61656303-39a3-4b92-9f2c-e5243a63ca21","Type":"ContainerDied","Data":"91b2cd7fe9f2ce0c16b6dc482f5066ccb6808acd40d0c9dec67e27558a212876"} Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.849488 4837 scope.go:117] "RemoveContainer" containerID="814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.849559 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swrvb" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.880937 4837 scope.go:117] "RemoveContainer" containerID="a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.909856 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.912899 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swrvb"] Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.919537 4837 scope.go:117] "RemoveContainer" containerID="c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.943378 4837 scope.go:117] "RemoveContainer" containerID="814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d" Oct 01 08:31:50 crc kubenswrapper[4837]: E1001 08:31:50.943942 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d\": container with ID starting with 814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d not found: ID does not exist" containerID="814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.943992 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d"} err="failed to get container status \"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d\": rpc error: code = NotFound desc = could not find container \"814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d\": container with ID starting with 814fc46cd070c1d58a1bc0760141c424b61413e2f06642af37ee94a1f1dd7f0d not found: ID does not exist" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.944077 4837 scope.go:117] "RemoveContainer" containerID="a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3" Oct 01 08:31:50 crc kubenswrapper[4837]: E1001 08:31:50.945354 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3\": container with ID starting with a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3 not found: ID does not exist" containerID="a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.945385 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3"} err="failed to get container status \"a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3\": rpc error: code = NotFound desc = could not find container \"a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3\": container with ID starting with a2d928e6636d4355ac7260b19072e4ffd8e0ed443b2e5f9c74311aa0e16a89d3 not found: ID does not exist" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.945404 4837 scope.go:117] "RemoveContainer" containerID="c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb" Oct 01 08:31:50 crc kubenswrapper[4837]: E1001 08:31:50.946651 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb\": container with ID starting with c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb not found: ID does not exist" containerID="c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb" Oct 01 08:31:50 crc kubenswrapper[4837]: I1001 08:31:50.946725 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb"} err="failed to get container status \"c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb\": rpc error: code = NotFound desc = could not find container \"c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb\": container with ID starting with c4c5f7e49b4224966f595112c93ee36716988716fd8d3c6e853b93fa909d96cb not found: ID does not exist" Oct 01 08:31:51 crc kubenswrapper[4837]: I1001 08:31:51.829222 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" path="/var/lib/kubelet/pods/61656303-39a3-4b92-9f2c-e5243a63ca21/volumes" Oct 01 08:31:52 crc kubenswrapper[4837]: I1001 08:31:52.826268 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:31:52 crc kubenswrapper[4837]: E1001 08:31:52.827027 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:32:07 crc kubenswrapper[4837]: I1001 08:32:07.816513 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:32:07 crc kubenswrapper[4837]: E1001 08:32:07.817446 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:32:19 crc kubenswrapper[4837]: I1001 08:32:19.816345 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:32:19 crc kubenswrapper[4837]: E1001 08:32:19.818192 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.555166 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:31 crc kubenswrapper[4837]: E1001 08:32:31.558492 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="extract-utilities" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.558659 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="extract-utilities" Oct 01 08:32:31 crc kubenswrapper[4837]: E1001 08:32:31.558836 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="registry-server" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.558956 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="registry-server" Oct 01 08:32:31 crc kubenswrapper[4837]: E1001 08:32:31.559074 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="extract-content" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.559192 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="extract-content" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.559579 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="61656303-39a3-4b92-9f2c-e5243a63ca21" containerName="registry-server" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.561500 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.570439 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.719494 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.719605 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-826cv\" (UniqueName: \"kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.719751 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.820851 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-826cv\" (UniqueName: \"kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.820906 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.820937 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.821835 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.822014 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.843780 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-826cv\" (UniqueName: \"kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv\") pod \"certified-operators-4xxvh\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:31 crc kubenswrapper[4837]: I1001 08:32:31.886896 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:32 crc kubenswrapper[4837]: I1001 08:32:32.341159 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:32 crc kubenswrapper[4837]: I1001 08:32:32.816396 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:32:32 crc kubenswrapper[4837]: E1001 08:32:32.817434 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:32:33 crc kubenswrapper[4837]: I1001 08:32:33.244183 4837 generic.go:334] "Generic (PLEG): container finished" podID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerID="2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e" exitCode=0 Oct 01 08:32:33 crc kubenswrapper[4837]: I1001 08:32:33.244255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerDied","Data":"2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e"} Oct 01 08:32:33 crc kubenswrapper[4837]: I1001 08:32:33.244306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerStarted","Data":"6394d48fdf5c4e1920b1abd090db4e7a875bed217872f57152df21e44a8cb98b"} Oct 01 08:32:35 crc kubenswrapper[4837]: I1001 08:32:35.266518 4837 generic.go:334] "Generic (PLEG): container finished" podID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerID="20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1" exitCode=0 Oct 01 08:32:35 crc kubenswrapper[4837]: I1001 08:32:35.266648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerDied","Data":"20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1"} Oct 01 08:32:37 crc kubenswrapper[4837]: I1001 08:32:37.289929 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerStarted","Data":"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e"} Oct 01 08:32:37 crc kubenswrapper[4837]: I1001 08:32:37.323303 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4xxvh" podStartSLOduration=3.253941194 podStartE2EDuration="6.323284103s" podCreationTimestamp="2025-10-01 08:32:31 +0000 UTC" firstStartedPulling="2025-10-01 08:32:33.247483425 +0000 UTC m=+5210.089090910" lastFinishedPulling="2025-10-01 08:32:36.316826344 +0000 UTC m=+5213.158433819" observedRunningTime="2025-10-01 08:32:37.317649705 +0000 UTC m=+5214.159257190" watchObservedRunningTime="2025-10-01 08:32:37.323284103 +0000 UTC m=+5214.164891548" Oct 01 08:32:41 crc kubenswrapper[4837]: I1001 08:32:41.887797 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:41 crc kubenswrapper[4837]: I1001 08:32:41.888280 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:41 crc kubenswrapper[4837]: I1001 08:32:41.940738 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:42 crc kubenswrapper[4837]: I1001 08:32:42.405467 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:42 crc kubenswrapper[4837]: I1001 08:32:42.466076 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.354874 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4xxvh" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="registry-server" containerID="cri-o://82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e" gracePeriod=2 Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.739549 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.816218 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:32:44 crc kubenswrapper[4837]: E1001 08:32:44.816612 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.818606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-826cv\" (UniqueName: \"kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv\") pod \"eae8545e-cfbd-44e8-bde4-34b95e256d00\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.818682 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content\") pod \"eae8545e-cfbd-44e8-bde4-34b95e256d00\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.818748 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities\") pod \"eae8545e-cfbd-44e8-bde4-34b95e256d00\" (UID: \"eae8545e-cfbd-44e8-bde4-34b95e256d00\") " Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.819915 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities" (OuterVolumeSpecName: "utilities") pod "eae8545e-cfbd-44e8-bde4-34b95e256d00" (UID: "eae8545e-cfbd-44e8-bde4-34b95e256d00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.825688 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv" (OuterVolumeSpecName: "kube-api-access-826cv") pod "eae8545e-cfbd-44e8-bde4-34b95e256d00" (UID: "eae8545e-cfbd-44e8-bde4-34b95e256d00"). InnerVolumeSpecName "kube-api-access-826cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.920660 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-826cv\" (UniqueName: \"kubernetes.io/projected/eae8545e-cfbd-44e8-bde4-34b95e256d00-kube-api-access-826cv\") on node \"crc\" DevicePath \"\"" Oct 01 08:32:44 crc kubenswrapper[4837]: I1001 08:32:44.920708 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.063469 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eae8545e-cfbd-44e8-bde4-34b95e256d00" (UID: "eae8545e-cfbd-44e8-bde4-34b95e256d00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.123343 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eae8545e-cfbd-44e8-bde4-34b95e256d00-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.368110 4837 generic.go:334] "Generic (PLEG): container finished" podID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerID="82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e" exitCode=0 Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.368179 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerDied","Data":"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e"} Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.368258 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4xxvh" event={"ID":"eae8545e-cfbd-44e8-bde4-34b95e256d00","Type":"ContainerDied","Data":"6394d48fdf5c4e1920b1abd090db4e7a875bed217872f57152df21e44a8cb98b"} Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.368282 4837 scope.go:117] "RemoveContainer" containerID="82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.368214 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4xxvh" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.400547 4837 scope.go:117] "RemoveContainer" containerID="20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.441380 4837 scope.go:117] "RemoveContainer" containerID="2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.460760 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.476274 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4xxvh"] Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.483843 4837 scope.go:117] "RemoveContainer" containerID="82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e" Oct 01 08:32:45 crc kubenswrapper[4837]: E1001 08:32:45.484804 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e\": container with ID starting with 82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e not found: ID does not exist" containerID="82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.484916 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e"} err="failed to get container status \"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e\": rpc error: code = NotFound desc = could not find container \"82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e\": container with ID starting with 82e2376567599e0330c2996131d5ef14fc81d8b700f0b2b1818037d99c49672e not found: ID does not exist" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.484973 4837 scope.go:117] "RemoveContainer" containerID="20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1" Oct 01 08:32:45 crc kubenswrapper[4837]: E1001 08:32:45.485512 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1\": container with ID starting with 20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1 not found: ID does not exist" containerID="20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.485611 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1"} err="failed to get container status \"20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1\": rpc error: code = NotFound desc = could not find container \"20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1\": container with ID starting with 20a088bc14ce779dbbf9100f591ffa3e6eaeb02527f0a1a3c67f8a28838d8ed1 not found: ID does not exist" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.485728 4837 scope.go:117] "RemoveContainer" containerID="2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e" Oct 01 08:32:45 crc kubenswrapper[4837]: E1001 08:32:45.486192 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e\": container with ID starting with 2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e not found: ID does not exist" containerID="2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.486225 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e"} err="failed to get container status \"2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e\": rpc error: code = NotFound desc = could not find container \"2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e\": container with ID starting with 2d1134d170318771636c66c1e7d06a817c9727c689a1a884276c155ac3f09f8e not found: ID does not exist" Oct 01 08:32:45 crc kubenswrapper[4837]: I1001 08:32:45.833363 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" path="/var/lib/kubelet/pods/eae8545e-cfbd-44e8-bde4-34b95e256d00/volumes" Oct 01 08:32:56 crc kubenswrapper[4837]: I1001 08:32:56.816404 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:32:56 crc kubenswrapper[4837]: E1001 08:32:56.817409 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:33:07 crc kubenswrapper[4837]: I1001 08:33:07.816461 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:33:07 crc kubenswrapper[4837]: E1001 08:33:07.817537 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:33:19 crc kubenswrapper[4837]: I1001 08:33:19.815842 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:33:19 crc kubenswrapper[4837]: E1001 08:33:19.816722 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:33:34 crc kubenswrapper[4837]: I1001 08:33:34.815902 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:33:35 crc kubenswrapper[4837]: I1001 08:33:35.827087 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f"} Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.466287 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:09 crc kubenswrapper[4837]: E1001 08:34:09.467388 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="extract-content" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.467410 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="extract-content" Oct 01 08:34:09 crc kubenswrapper[4837]: E1001 08:34:09.467464 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="registry-server" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.467478 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="registry-server" Oct 01 08:34:09 crc kubenswrapper[4837]: E1001 08:34:09.467507 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="extract-utilities" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.467521 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="extract-utilities" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.467820 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae8545e-cfbd-44e8-bde4-34b95e256d00" containerName="registry-server" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.469633 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.476626 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.600902 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.600969 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvthl\" (UniqueName: \"kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.601196 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.702751 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvthl\" (UniqueName: \"kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.703166 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.703640 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.703753 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.704071 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.729848 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvthl\" (UniqueName: \"kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl\") pod \"redhat-operators-qtcrs\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:09 crc kubenswrapper[4837]: I1001 08:34:09.806631 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:10 crc kubenswrapper[4837]: I1001 08:34:10.267121 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:11 crc kubenswrapper[4837]: I1001 08:34:11.130940 4837 generic.go:334] "Generic (PLEG): container finished" podID="265687f2-8c2a-465d-86af-09c7b8e95634" containerID="273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd" exitCode=0 Oct 01 08:34:11 crc kubenswrapper[4837]: I1001 08:34:11.131023 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerDied","Data":"273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd"} Oct 01 08:34:11 crc kubenswrapper[4837]: I1001 08:34:11.131102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerStarted","Data":"1fe34e73fdb01fabb449a9e5079fec040a76a9797a0afafe7dcedd4333e867d0"} Oct 01 08:34:12 crc kubenswrapper[4837]: I1001 08:34:12.148478 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerStarted","Data":"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab"} Oct 01 08:34:13 crc kubenswrapper[4837]: I1001 08:34:13.163387 4837 generic.go:334] "Generic (PLEG): container finished" podID="265687f2-8c2a-465d-86af-09c7b8e95634" containerID="eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab" exitCode=0 Oct 01 08:34:13 crc kubenswrapper[4837]: I1001 08:34:13.163511 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerDied","Data":"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab"} Oct 01 08:34:14 crc kubenswrapper[4837]: I1001 08:34:14.176042 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerStarted","Data":"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57"} Oct 01 08:34:14 crc kubenswrapper[4837]: I1001 08:34:14.201856 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qtcrs" podStartSLOduration=2.601775387 podStartE2EDuration="5.20182924s" podCreationTimestamp="2025-10-01 08:34:09 +0000 UTC" firstStartedPulling="2025-10-01 08:34:11.133837952 +0000 UTC m=+5307.975445447" lastFinishedPulling="2025-10-01 08:34:13.733891805 +0000 UTC m=+5310.575499300" observedRunningTime="2025-10-01 08:34:14.200431636 +0000 UTC m=+5311.042039111" watchObservedRunningTime="2025-10-01 08:34:14.20182924 +0000 UTC m=+5311.043436715" Oct 01 08:34:19 crc kubenswrapper[4837]: I1001 08:34:19.807123 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:19 crc kubenswrapper[4837]: I1001 08:34:19.807661 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:19 crc kubenswrapper[4837]: I1001 08:34:19.856816 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:20 crc kubenswrapper[4837]: I1001 08:34:20.300379 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:20 crc kubenswrapper[4837]: I1001 08:34:20.365934 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.250213 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qtcrs" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="registry-server" containerID="cri-o://debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57" gracePeriod=2 Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.751272 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.908439 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities\") pod \"265687f2-8c2a-465d-86af-09c7b8e95634\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.908524 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content\") pod \"265687f2-8c2a-465d-86af-09c7b8e95634\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.908645 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvthl\" (UniqueName: \"kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl\") pod \"265687f2-8c2a-465d-86af-09c7b8e95634\" (UID: \"265687f2-8c2a-465d-86af-09c7b8e95634\") " Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.910104 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities" (OuterVolumeSpecName: "utilities") pod "265687f2-8c2a-465d-86af-09c7b8e95634" (UID: "265687f2-8c2a-465d-86af-09c7b8e95634"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:34:22 crc kubenswrapper[4837]: I1001 08:34:22.931226 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl" (OuterVolumeSpecName: "kube-api-access-rvthl") pod "265687f2-8c2a-465d-86af-09c7b8e95634" (UID: "265687f2-8c2a-465d-86af-09c7b8e95634"). InnerVolumeSpecName "kube-api-access-rvthl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.010759 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.010821 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvthl\" (UniqueName: \"kubernetes.io/projected/265687f2-8c2a-465d-86af-09c7b8e95634-kube-api-access-rvthl\") on node \"crc\" DevicePath \"\"" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.263352 4837 generic.go:334] "Generic (PLEG): container finished" podID="265687f2-8c2a-465d-86af-09c7b8e95634" containerID="debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57" exitCode=0 Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.263424 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerDied","Data":"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57"} Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.263466 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtcrs" event={"ID":"265687f2-8c2a-465d-86af-09c7b8e95634","Type":"ContainerDied","Data":"1fe34e73fdb01fabb449a9e5079fec040a76a9797a0afafe7dcedd4333e867d0"} Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.263498 4837 scope.go:117] "RemoveContainer" containerID="debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.263684 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtcrs" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.299331 4837 scope.go:117] "RemoveContainer" containerID="eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.329755 4837 scope.go:117] "RemoveContainer" containerID="273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.387348 4837 scope.go:117] "RemoveContainer" containerID="debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57" Oct 01 08:34:23 crc kubenswrapper[4837]: E1001 08:34:23.388595 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57\": container with ID starting with debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57 not found: ID does not exist" containerID="debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.388636 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57"} err="failed to get container status \"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57\": rpc error: code = NotFound desc = could not find container \"debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57\": container with ID starting with debe91cc177060655e2f8b56c3ac1ad4f134d92aaf24cad270ab966826ddee57 not found: ID does not exist" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.388735 4837 scope.go:117] "RemoveContainer" containerID="eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab" Oct 01 08:34:23 crc kubenswrapper[4837]: E1001 08:34:23.389337 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab\": container with ID starting with eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab not found: ID does not exist" containerID="eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.390430 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab"} err="failed to get container status \"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab\": rpc error: code = NotFound desc = could not find container \"eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab\": container with ID starting with eacc9240c42a81c785d6bc741b399b769661d6003c668726e6ef5e4a8badd8ab not found: ID does not exist" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.390541 4837 scope.go:117] "RemoveContainer" containerID="273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd" Oct 01 08:34:23 crc kubenswrapper[4837]: E1001 08:34:23.391478 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd\": container with ID starting with 273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd not found: ID does not exist" containerID="273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.391548 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd"} err="failed to get container status \"273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd\": rpc error: code = NotFound desc = could not find container \"273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd\": container with ID starting with 273d98f9f19f99eefb0ceef62d3970b62d2c39f93bd788d0e412282f1d1873bd not found: ID does not exist" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.616082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "265687f2-8c2a-465d-86af-09c7b8e95634" (UID: "265687f2-8c2a-465d-86af-09c7b8e95634"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.625347 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/265687f2-8c2a-465d-86af-09c7b8e95634-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.904997 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:23 crc kubenswrapper[4837]: I1001 08:34:23.917746 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qtcrs"] Oct 01 08:34:25 crc kubenswrapper[4837]: I1001 08:34:25.828939 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" path="/var/lib/kubelet/pods/265687f2-8c2a-465d-86af-09c7b8e95634/volumes" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.126879 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:00 crc kubenswrapper[4837]: E1001 08:35:00.127793 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="extract-content" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.127809 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="extract-content" Oct 01 08:35:00 crc kubenswrapper[4837]: E1001 08:35:00.127826 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="registry-server" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.127837 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="registry-server" Oct 01 08:35:00 crc kubenswrapper[4837]: E1001 08:35:00.127855 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="extract-utilities" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.127863 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="extract-utilities" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.128045 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="265687f2-8c2a-465d-86af-09c7b8e95634" containerName="registry-server" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.129374 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.134930 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.284707 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pwrv\" (UniqueName: \"kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.284791 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.284824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.385946 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.386013 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.386053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pwrv\" (UniqueName: \"kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.386503 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.386598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.417821 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pwrv\" (UniqueName: \"kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv\") pod \"redhat-marketplace-nn6kr\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.450620 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:00 crc kubenswrapper[4837]: I1001 08:35:00.929177 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:00 crc kubenswrapper[4837]: W1001 08:35:00.945945 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cadbc36_14da_4ab5_80f9_3fd27b798d10.slice/crio-3699f03379a7860b5cc1e4f63224c9556d3df8d5afd6afc3bdd16391a3ffc058 WatchSource:0}: Error finding container 3699f03379a7860b5cc1e4f63224c9556d3df8d5afd6afc3bdd16391a3ffc058: Status 404 returned error can't find the container with id 3699f03379a7860b5cc1e4f63224c9556d3df8d5afd6afc3bdd16391a3ffc058 Oct 01 08:35:01 crc kubenswrapper[4837]: I1001 08:35:01.654665 4837 generic.go:334] "Generic (PLEG): container finished" podID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerID="5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573" exitCode=0 Oct 01 08:35:01 crc kubenswrapper[4837]: I1001 08:35:01.654795 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerDied","Data":"5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573"} Oct 01 08:35:01 crc kubenswrapper[4837]: I1001 08:35:01.655092 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerStarted","Data":"3699f03379a7860b5cc1e4f63224c9556d3df8d5afd6afc3bdd16391a3ffc058"} Oct 01 08:35:02 crc kubenswrapper[4837]: I1001 08:35:02.669664 4837 generic.go:334] "Generic (PLEG): container finished" podID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerID="47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412" exitCode=0 Oct 01 08:35:02 crc kubenswrapper[4837]: I1001 08:35:02.669909 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerDied","Data":"47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412"} Oct 01 08:35:03 crc kubenswrapper[4837]: I1001 08:35:03.679918 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerStarted","Data":"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77"} Oct 01 08:35:03 crc kubenswrapper[4837]: I1001 08:35:03.701440 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nn6kr" podStartSLOduration=2.156388979 podStartE2EDuration="3.701422296s" podCreationTimestamp="2025-10-01 08:35:00 +0000 UTC" firstStartedPulling="2025-10-01 08:35:01.658835667 +0000 UTC m=+5358.500443152" lastFinishedPulling="2025-10-01 08:35:03.203869014 +0000 UTC m=+5360.045476469" observedRunningTime="2025-10-01 08:35:03.697753525 +0000 UTC m=+5360.539360990" watchObservedRunningTime="2025-10-01 08:35:03.701422296 +0000 UTC m=+5360.543029761" Oct 01 08:35:10 crc kubenswrapper[4837]: I1001 08:35:10.450933 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:10 crc kubenswrapper[4837]: I1001 08:35:10.451571 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:10 crc kubenswrapper[4837]: I1001 08:35:10.532815 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:10 crc kubenswrapper[4837]: I1001 08:35:10.803053 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:10 crc kubenswrapper[4837]: I1001 08:35:10.844479 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:12 crc kubenswrapper[4837]: I1001 08:35:12.763271 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nn6kr" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="registry-server" containerID="cri-o://5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77" gracePeriod=2 Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.228746 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.394752 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content\") pod \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.395178 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pwrv\" (UniqueName: \"kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv\") pod \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.395344 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities\") pod \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\" (UID: \"1cadbc36-14da-4ab5-80f9-3fd27b798d10\") " Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.396804 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities" (OuterVolumeSpecName: "utilities") pod "1cadbc36-14da-4ab5-80f9-3fd27b798d10" (UID: "1cadbc36-14da-4ab5-80f9-3fd27b798d10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.403924 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv" (OuterVolumeSpecName: "kube-api-access-6pwrv") pod "1cadbc36-14da-4ab5-80f9-3fd27b798d10" (UID: "1cadbc36-14da-4ab5-80f9-3fd27b798d10"). InnerVolumeSpecName "kube-api-access-6pwrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.411303 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1cadbc36-14da-4ab5-80f9-3fd27b798d10" (UID: "1cadbc36-14da-4ab5-80f9-3fd27b798d10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.497376 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.497424 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cadbc36-14da-4ab5-80f9-3fd27b798d10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.497473 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pwrv\" (UniqueName: \"kubernetes.io/projected/1cadbc36-14da-4ab5-80f9-3fd27b798d10-kube-api-access-6pwrv\") on node \"crc\" DevicePath \"\"" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.775588 4837 generic.go:334] "Generic (PLEG): container finished" podID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerID="5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77" exitCode=0 Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.775652 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerDied","Data":"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77"} Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.775777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nn6kr" event={"ID":"1cadbc36-14da-4ab5-80f9-3fd27b798d10","Type":"ContainerDied","Data":"3699f03379a7860b5cc1e4f63224c9556d3df8d5afd6afc3bdd16391a3ffc058"} Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.775816 4837 scope.go:117] "RemoveContainer" containerID="5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.775635 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nn6kr" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.820338 4837 scope.go:117] "RemoveContainer" containerID="47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.835140 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.835191 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nn6kr"] Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.855139 4837 scope.go:117] "RemoveContainer" containerID="5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.896833 4837 scope.go:117] "RemoveContainer" containerID="5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77" Oct 01 08:35:13 crc kubenswrapper[4837]: E1001 08:35:13.897352 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77\": container with ID starting with 5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77 not found: ID does not exist" containerID="5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.897407 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77"} err="failed to get container status \"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77\": rpc error: code = NotFound desc = could not find container \"5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77\": container with ID starting with 5394d34e0588804171ed6a271f1d0f54c5b2d36ef1a86175ba33f11f96f91f77 not found: ID does not exist" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.897441 4837 scope.go:117] "RemoveContainer" containerID="47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412" Oct 01 08:35:13 crc kubenswrapper[4837]: E1001 08:35:13.897981 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412\": container with ID starting with 47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412 not found: ID does not exist" containerID="47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.898047 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412"} err="failed to get container status \"47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412\": rpc error: code = NotFound desc = could not find container \"47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412\": container with ID starting with 47d1c5c28b713bf8d3e60cdfd93158526bbf68f8c63630d8924355cdc5656412 not found: ID does not exist" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.898090 4837 scope.go:117] "RemoveContainer" containerID="5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573" Oct 01 08:35:13 crc kubenswrapper[4837]: E1001 08:35:13.898434 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573\": container with ID starting with 5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573 not found: ID does not exist" containerID="5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573" Oct 01 08:35:13 crc kubenswrapper[4837]: I1001 08:35:13.898463 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573"} err="failed to get container status \"5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573\": rpc error: code = NotFound desc = could not find container \"5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573\": container with ID starting with 5b1bc3aba3ed3881cf0f207e6bf3fb2c178b7ab7d391cd5e3d3ae166e6a06573 not found: ID does not exist" Oct 01 08:35:15 crc kubenswrapper[4837]: I1001 08:35:15.834900 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" path="/var/lib/kubelet/pods/1cadbc36-14da-4ab5-80f9-3fd27b798d10/volumes" Oct 01 08:35:53 crc kubenswrapper[4837]: I1001 08:35:53.079621 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:35:53 crc kubenswrapper[4837]: I1001 08:35:53.080445 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:36:23 crc kubenswrapper[4837]: I1001 08:36:23.080009 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:36:23 crc kubenswrapper[4837]: I1001 08:36:23.080825 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.079668 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.080540 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.080633 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.081885 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.081994 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f" gracePeriod=600 Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.741304 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f" exitCode=0 Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.741403 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f"} Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.741770 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088"} Oct 01 08:36:53 crc kubenswrapper[4837]: I1001 08:36:53.741799 4837 scope.go:117] "RemoveContainer" containerID="3dd2c9b74fabfa20955907c9f171e656f9015e630a84f408bdecd78cb80e5f63" Oct 01 08:38:53 crc kubenswrapper[4837]: I1001 08:38:53.079758 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:38:53 crc kubenswrapper[4837]: I1001 08:38:53.080717 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:39:23 crc kubenswrapper[4837]: I1001 08:39:23.079551 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:39:23 crc kubenswrapper[4837]: I1001 08:39:23.080326 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.079594 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.080168 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.080223 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.080933 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.080994 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" gracePeriod=600 Oct 01 08:39:53 crc kubenswrapper[4837]: E1001 08:39:53.209722 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.449783 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" exitCode=0 Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.449851 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088"} Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.449914 4837 scope.go:117] "RemoveContainer" containerID="425ea447a92814051651b75d433de07adfa49593aafdadcfd3c178421e2efe7f" Oct 01 08:39:53 crc kubenswrapper[4837]: I1001 08:39:53.450987 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:39:53 crc kubenswrapper[4837]: E1001 08:39:53.451782 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:40:04 crc kubenswrapper[4837]: I1001 08:40:04.816362 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:40:04 crc kubenswrapper[4837]: E1001 08:40:04.817749 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:40:16 crc kubenswrapper[4837]: I1001 08:40:16.815413 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:40:16 crc kubenswrapper[4837]: E1001 08:40:16.816229 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:40:30 crc kubenswrapper[4837]: I1001 08:40:30.816430 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:40:30 crc kubenswrapper[4837]: E1001 08:40:30.817379 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:40:45 crc kubenswrapper[4837]: I1001 08:40:45.816725 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:40:45 crc kubenswrapper[4837]: E1001 08:40:45.817815 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:40:57 crc kubenswrapper[4837]: I1001 08:40:57.816468 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:40:57 crc kubenswrapper[4837]: E1001 08:40:57.817674 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:41:09 crc kubenswrapper[4837]: I1001 08:41:09.838598 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:41:09 crc kubenswrapper[4837]: E1001 08:41:09.839571 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:41:20 crc kubenswrapper[4837]: I1001 08:41:20.815815 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:41:20 crc kubenswrapper[4837]: E1001 08:41:20.816568 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:41:33 crc kubenswrapper[4837]: I1001 08:41:33.827603 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:41:33 crc kubenswrapper[4837]: E1001 08:41:33.828647 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.633724 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:36 crc kubenswrapper[4837]: E1001 08:41:36.634414 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="extract-content" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.634427 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="extract-content" Oct 01 08:41:36 crc kubenswrapper[4837]: E1001 08:41:36.634450 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="extract-utilities" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.634456 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="extract-utilities" Oct 01 08:41:36 crc kubenswrapper[4837]: E1001 08:41:36.634467 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="registry-server" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.634474 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="registry-server" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.634594 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cadbc36-14da-4ab5-80f9-3fd27b798d10" containerName="registry-server" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.635560 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.645314 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.728582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4whq\" (UniqueName: \"kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.728655 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.728806 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.831154 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4whq\" (UniqueName: \"kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.831260 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.831300 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.831902 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.832257 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.859841 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4whq\" (UniqueName: \"kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq\") pod \"community-operators-mwb9h\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:36 crc kubenswrapper[4837]: I1001 08:41:36.966386 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:37 crc kubenswrapper[4837]: I1001 08:41:37.513465 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:38 crc kubenswrapper[4837]: I1001 08:41:38.514646 4837 generic.go:334] "Generic (PLEG): container finished" podID="d28da709-7a94-4543-a546-17f63644113d" containerID="6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df" exitCode=0 Oct 01 08:41:38 crc kubenswrapper[4837]: I1001 08:41:38.514754 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerDied","Data":"6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df"} Oct 01 08:41:38 crc kubenswrapper[4837]: I1001 08:41:38.515076 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerStarted","Data":"28f60f92e12326d52273097e5384e1aa27bdae13e0bce47a028c65588ac67beb"} Oct 01 08:41:38 crc kubenswrapper[4837]: I1001 08:41:38.517483 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:41:40 crc kubenswrapper[4837]: I1001 08:41:40.535862 4837 generic.go:334] "Generic (PLEG): container finished" podID="d28da709-7a94-4543-a546-17f63644113d" containerID="889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced" exitCode=0 Oct 01 08:41:40 crc kubenswrapper[4837]: I1001 08:41:40.535964 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerDied","Data":"889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced"} Oct 01 08:41:41 crc kubenswrapper[4837]: I1001 08:41:41.547003 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerStarted","Data":"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac"} Oct 01 08:41:41 crc kubenswrapper[4837]: I1001 08:41:41.573140 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mwb9h" podStartSLOduration=2.882551479 podStartE2EDuration="5.573110467s" podCreationTimestamp="2025-10-01 08:41:36 +0000 UTC" firstStartedPulling="2025-10-01 08:41:38.517084584 +0000 UTC m=+5755.358692039" lastFinishedPulling="2025-10-01 08:41:41.207643542 +0000 UTC m=+5758.049251027" observedRunningTime="2025-10-01 08:41:41.567620912 +0000 UTC m=+5758.409228367" watchObservedRunningTime="2025-10-01 08:41:41.573110467 +0000 UTC m=+5758.414717932" Oct 01 08:41:46 crc kubenswrapper[4837]: I1001 08:41:46.816850 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:41:46 crc kubenswrapper[4837]: E1001 08:41:46.820051 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:41:46 crc kubenswrapper[4837]: I1001 08:41:46.967072 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:46 crc kubenswrapper[4837]: I1001 08:41:46.967127 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:47 crc kubenswrapper[4837]: I1001 08:41:47.012914 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:47 crc kubenswrapper[4837]: I1001 08:41:47.649032 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:52 crc kubenswrapper[4837]: I1001 08:41:52.586139 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:52 crc kubenswrapper[4837]: I1001 08:41:52.586956 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mwb9h" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="registry-server" containerID="cri-o://eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac" gracePeriod=2 Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.123327 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.231097 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4whq\" (UniqueName: \"kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq\") pod \"d28da709-7a94-4543-a546-17f63644113d\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.231294 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities\") pod \"d28da709-7a94-4543-a546-17f63644113d\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.231347 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content\") pod \"d28da709-7a94-4543-a546-17f63644113d\" (UID: \"d28da709-7a94-4543-a546-17f63644113d\") " Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.232716 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities" (OuterVolumeSpecName: "utilities") pod "d28da709-7a94-4543-a546-17f63644113d" (UID: "d28da709-7a94-4543-a546-17f63644113d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.237878 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq" (OuterVolumeSpecName: "kube-api-access-c4whq") pod "d28da709-7a94-4543-a546-17f63644113d" (UID: "d28da709-7a94-4543-a546-17f63644113d"). InnerVolumeSpecName "kube-api-access-c4whq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.288861 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d28da709-7a94-4543-a546-17f63644113d" (UID: "d28da709-7a94-4543-a546-17f63644113d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.333903 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4whq\" (UniqueName: \"kubernetes.io/projected/d28da709-7a94-4543-a546-17f63644113d-kube-api-access-c4whq\") on node \"crc\" DevicePath \"\"" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.333951 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.333965 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28da709-7a94-4543-a546-17f63644113d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.663574 4837 generic.go:334] "Generic (PLEG): container finished" podID="d28da709-7a94-4543-a546-17f63644113d" containerID="eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac" exitCode=0 Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.663753 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwb9h" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.663765 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerDied","Data":"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac"} Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.664224 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwb9h" event={"ID":"d28da709-7a94-4543-a546-17f63644113d","Type":"ContainerDied","Data":"28f60f92e12326d52273097e5384e1aa27bdae13e0bce47a028c65588ac67beb"} Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.664265 4837 scope.go:117] "RemoveContainer" containerID="eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.702539 4837 scope.go:117] "RemoveContainer" containerID="889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.707024 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.721084 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mwb9h"] Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.736644 4837 scope.go:117] "RemoveContainer" containerID="6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.771581 4837 scope.go:117] "RemoveContainer" containerID="eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac" Oct 01 08:41:53 crc kubenswrapper[4837]: E1001 08:41:53.772634 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac\": container with ID starting with eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac not found: ID does not exist" containerID="eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.772885 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac"} err="failed to get container status \"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac\": rpc error: code = NotFound desc = could not find container \"eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac\": container with ID starting with eae0b16f2d3b667ab6d2539aa3e626a64359559a4e3d62c4f657c287e4dfbcac not found: ID does not exist" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.773013 4837 scope.go:117] "RemoveContainer" containerID="889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced" Oct 01 08:41:53 crc kubenswrapper[4837]: E1001 08:41:53.774887 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced\": container with ID starting with 889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced not found: ID does not exist" containerID="889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.774928 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced"} err="failed to get container status \"889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced\": rpc error: code = NotFound desc = could not find container \"889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced\": container with ID starting with 889cdb7c39d920a8b08c52cb89da9d3948544d2034124c2b10c47db9a2b16ced not found: ID does not exist" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.774960 4837 scope.go:117] "RemoveContainer" containerID="6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df" Oct 01 08:41:53 crc kubenswrapper[4837]: E1001 08:41:53.775262 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df\": container with ID starting with 6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df not found: ID does not exist" containerID="6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.775290 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df"} err="failed to get container status \"6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df\": rpc error: code = NotFound desc = could not find container \"6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df\": container with ID starting with 6f6f579f40ffea1f7d7ff154815d8e720ecdd663283a884b833fd5d206b2f9df not found: ID does not exist" Oct 01 08:41:53 crc kubenswrapper[4837]: I1001 08:41:53.833557 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d28da709-7a94-4543-a546-17f63644113d" path="/var/lib/kubelet/pods/d28da709-7a94-4543-a546-17f63644113d/volumes" Oct 01 08:42:00 crc kubenswrapper[4837]: I1001 08:42:00.817049 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:42:00 crc kubenswrapper[4837]: E1001 08:42:00.819913 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:42:15 crc kubenswrapper[4837]: I1001 08:42:15.816624 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:42:15 crc kubenswrapper[4837]: E1001 08:42:15.817658 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:42:30 crc kubenswrapper[4837]: I1001 08:42:30.816235 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:42:30 crc kubenswrapper[4837]: E1001 08:42:30.817445 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:42:44 crc kubenswrapper[4837]: I1001 08:42:44.816013 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:42:44 crc kubenswrapper[4837]: E1001 08:42:44.817066 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:42:58 crc kubenswrapper[4837]: I1001 08:42:58.816259 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:42:58 crc kubenswrapper[4837]: E1001 08:42:58.817194 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:43:11 crc kubenswrapper[4837]: I1001 08:43:11.816413 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:43:11 crc kubenswrapper[4837]: E1001 08:43:11.817195 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:43:26 crc kubenswrapper[4837]: I1001 08:43:26.816072 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:43:26 crc kubenswrapper[4837]: E1001 08:43:26.817322 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:43:41 crc kubenswrapper[4837]: I1001 08:43:41.816860 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:43:41 crc kubenswrapper[4837]: E1001 08:43:41.817883 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.043828 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-4fp78"] Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.058554 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-4fp78"] Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.160788 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-s4g8k"] Oct 01 08:43:49 crc kubenswrapper[4837]: E1001 08:43:49.161499 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="extract-content" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.161550 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="extract-content" Oct 01 08:43:49 crc kubenswrapper[4837]: E1001 08:43:49.161594 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="registry-server" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.161611 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="registry-server" Oct 01 08:43:49 crc kubenswrapper[4837]: E1001 08:43:49.161654 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="extract-utilities" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.161672 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="extract-utilities" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.162055 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28da709-7a94-4543-a546-17f63644113d" containerName="registry-server" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.163100 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.169144 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.169228 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.169403 4837 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5xt95" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.172492 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.175965 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-s4g8k"] Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.309850 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvbs\" (UniqueName: \"kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.310321 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.310383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.413466 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvbs\" (UniqueName: \"kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.413663 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.413761 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.414449 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.416652 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.447712 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvbs\" (UniqueName: \"kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs\") pod \"crc-storage-crc-s4g8k\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.498869 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:49 crc kubenswrapper[4837]: I1001 08:43:49.828412 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff67284-dd9c-485e-bb92-38e108c5d2af" path="/var/lib/kubelet/pods/eff67284-dd9c-485e-bb92-38e108c5d2af/volumes" Oct 01 08:43:50 crc kubenswrapper[4837]: I1001 08:43:50.108328 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-s4g8k"] Oct 01 08:43:50 crc kubenswrapper[4837]: I1001 08:43:50.787237 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4g8k" event={"ID":"526bf653-203b-4bb0-b441-b81eef11ef04","Type":"ContainerStarted","Data":"99a9161f2426bef4079b47946463b9016d0f72ab811b8300603a5a37a51f7de8"} Oct 01 08:43:51 crc kubenswrapper[4837]: I1001 08:43:51.798574 4837 generic.go:334] "Generic (PLEG): container finished" podID="526bf653-203b-4bb0-b441-b81eef11ef04" containerID="9eaf6a53d9ce56cab43fb1e811b1916ba25dd9eac1509d9536edd5f6f601e89c" exitCode=0 Oct 01 08:43:51 crc kubenswrapper[4837]: I1001 08:43:51.798625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4g8k" event={"ID":"526bf653-203b-4bb0-b441-b81eef11ef04","Type":"ContainerDied","Data":"9eaf6a53d9ce56cab43fb1e811b1916ba25dd9eac1509d9536edd5f6f601e89c"} Oct 01 08:43:52 crc kubenswrapper[4837]: I1001 08:43:52.816535 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:43:52 crc kubenswrapper[4837]: E1001 08:43:52.817159 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.156260 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.274888 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage\") pod \"526bf653-203b-4bb0-b441-b81eef11ef04\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.275651 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt\") pod \"526bf653-203b-4bb0-b441-b81eef11ef04\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.275716 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbvbs\" (UniqueName: \"kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs\") pod \"526bf653-203b-4bb0-b441-b81eef11ef04\" (UID: \"526bf653-203b-4bb0-b441-b81eef11ef04\") " Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.275782 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "526bf653-203b-4bb0-b441-b81eef11ef04" (UID: "526bf653-203b-4bb0-b441-b81eef11ef04"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.276084 4837 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/526bf653-203b-4bb0-b441-b81eef11ef04-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.288101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs" (OuterVolumeSpecName: "kube-api-access-rbvbs") pod "526bf653-203b-4bb0-b441-b81eef11ef04" (UID: "526bf653-203b-4bb0-b441-b81eef11ef04"). InnerVolumeSpecName "kube-api-access-rbvbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.309540 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "526bf653-203b-4bb0-b441-b81eef11ef04" (UID: "526bf653-203b-4bb0-b441-b81eef11ef04"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.377614 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbvbs\" (UniqueName: \"kubernetes.io/projected/526bf653-203b-4bb0-b441-b81eef11ef04-kube-api-access-rbvbs\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.377660 4837 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/526bf653-203b-4bb0-b441-b81eef11ef04-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.822460 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-s4g8k" Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.827466 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-s4g8k" event={"ID":"526bf653-203b-4bb0-b441-b81eef11ef04","Type":"ContainerDied","Data":"99a9161f2426bef4079b47946463b9016d0f72ab811b8300603a5a37a51f7de8"} Oct 01 08:43:53 crc kubenswrapper[4837]: I1001 08:43:53.827538 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99a9161f2426bef4079b47946463b9016d0f72ab811b8300603a5a37a51f7de8" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.389272 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-s4g8k"] Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.397810 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-s4g8k"] Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.519534 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-4fzw2"] Oct 01 08:43:55 crc kubenswrapper[4837]: E1001 08:43:55.520036 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526bf653-203b-4bb0-b441-b81eef11ef04" containerName="storage" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.520070 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="526bf653-203b-4bb0-b441-b81eef11ef04" containerName="storage" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.520364 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="526bf653-203b-4bb0-b441-b81eef11ef04" containerName="storage" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.521280 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.529843 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.530013 4837 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5xt95" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.530266 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.530369 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.534261 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4fzw2"] Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.612731 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.612779 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.612843 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cspqc\" (UniqueName: \"kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.714597 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.714659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.714759 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cspqc\" (UniqueName: \"kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.715163 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.715601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.751980 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cspqc\" (UniqueName: \"kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc\") pod \"crc-storage-crc-4fzw2\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.835868 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="526bf653-203b-4bb0-b441-b81eef11ef04" path="/var/lib/kubelet/pods/526bf653-203b-4bb0-b441-b81eef11ef04/volumes" Oct 01 08:43:55 crc kubenswrapper[4837]: I1001 08:43:55.844230 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:56 crc kubenswrapper[4837]: I1001 08:43:56.113617 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4fzw2"] Oct 01 08:43:56 crc kubenswrapper[4837]: I1001 08:43:56.850747 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fzw2" event={"ID":"964ab6ed-1d5c-489b-9251-529e25875d11","Type":"ContainerStarted","Data":"5e9af6faf263df862762679aa75b36a19b62162b8e352c8c600bcc3a7e48f1aa"} Oct 01 08:43:57 crc kubenswrapper[4837]: I1001 08:43:57.872711 4837 generic.go:334] "Generic (PLEG): container finished" podID="964ab6ed-1d5c-489b-9251-529e25875d11" containerID="e5ca74571e63e9f19cf260b6881139d586c1b679e70e7caf5d09573587b954da" exitCode=0 Oct 01 08:43:57 crc kubenswrapper[4837]: I1001 08:43:57.872803 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fzw2" event={"ID":"964ab6ed-1d5c-489b-9251-529e25875d11","Type":"ContainerDied","Data":"e5ca74571e63e9f19cf260b6881139d586c1b679e70e7caf5d09573587b954da"} Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.290959 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.379648 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage\") pod \"964ab6ed-1d5c-489b-9251-529e25875d11\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.379913 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cspqc\" (UniqueName: \"kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc\") pod \"964ab6ed-1d5c-489b-9251-529e25875d11\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.379975 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt\") pod \"964ab6ed-1d5c-489b-9251-529e25875d11\" (UID: \"964ab6ed-1d5c-489b-9251-529e25875d11\") " Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.380240 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "964ab6ed-1d5c-489b-9251-529e25875d11" (UID: "964ab6ed-1d5c-489b-9251-529e25875d11"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.380531 4837 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/964ab6ed-1d5c-489b-9251-529e25875d11-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.388018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc" (OuterVolumeSpecName: "kube-api-access-cspqc") pod "964ab6ed-1d5c-489b-9251-529e25875d11" (UID: "964ab6ed-1d5c-489b-9251-529e25875d11"). InnerVolumeSpecName "kube-api-access-cspqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.412390 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "964ab6ed-1d5c-489b-9251-529e25875d11" (UID: "964ab6ed-1d5c-489b-9251-529e25875d11"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.482188 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cspqc\" (UniqueName: \"kubernetes.io/projected/964ab6ed-1d5c-489b-9251-529e25875d11-kube-api-access-cspqc\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.482239 4837 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/964ab6ed-1d5c-489b-9251-529e25875d11-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.904460 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4fzw2" event={"ID":"964ab6ed-1d5c-489b-9251-529e25875d11","Type":"ContainerDied","Data":"5e9af6faf263df862762679aa75b36a19b62162b8e352c8c600bcc3a7e48f1aa"} Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.905010 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e9af6faf263df862762679aa75b36a19b62162b8e352c8c600bcc3a7e48f1aa" Oct 01 08:43:59 crc kubenswrapper[4837]: I1001 08:43:59.904543 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4fzw2" Oct 01 08:44:04 crc kubenswrapper[4837]: I1001 08:44:04.817084 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:44:04 crc kubenswrapper[4837]: E1001 08:44:04.818265 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:44:05 crc kubenswrapper[4837]: I1001 08:44:05.312350 4837 scope.go:117] "RemoveContainer" containerID="03b8114fbe7ef55b0cc69022bd642af7495d794afe4e71e7f326201ca1e32c42" Oct 01 08:44:17 crc kubenswrapper[4837]: I1001 08:44:17.815988 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:44:17 crc kubenswrapper[4837]: E1001 08:44:17.817059 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.230394 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:19 crc kubenswrapper[4837]: E1001 08:44:19.231117 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964ab6ed-1d5c-489b-9251-529e25875d11" containerName="storage" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.231133 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="964ab6ed-1d5c-489b-9251-529e25875d11" containerName="storage" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.231285 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="964ab6ed-1d5c-489b-9251-529e25875d11" containerName="storage" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.232466 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.261107 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.429296 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.429380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjzjl\" (UniqueName: \"kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.429423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.531246 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjzjl\" (UniqueName: \"kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.531318 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.531363 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.531866 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.532078 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.550528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjzjl\" (UniqueName: \"kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl\") pod \"redhat-operators-r9jnn\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:19 crc kubenswrapper[4837]: I1001 08:44:19.586708 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:20 crc kubenswrapper[4837]: I1001 08:44:20.082441 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:20 crc kubenswrapper[4837]: I1001 08:44:20.110562 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerStarted","Data":"9b05457a4528a0f3918c492b52f958d762eb805a40eb71f7c78c428217cb7612"} Oct 01 08:44:21 crc kubenswrapper[4837]: I1001 08:44:21.131049 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerID="d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd" exitCode=0 Oct 01 08:44:21 crc kubenswrapper[4837]: I1001 08:44:21.131118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerDied","Data":"d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd"} Oct 01 08:44:23 crc kubenswrapper[4837]: I1001 08:44:23.153288 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerID="ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508" exitCode=0 Oct 01 08:44:23 crc kubenswrapper[4837]: I1001 08:44:23.153600 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerDied","Data":"ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508"} Oct 01 08:44:24 crc kubenswrapper[4837]: I1001 08:44:24.166407 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerStarted","Data":"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae"} Oct 01 08:44:24 crc kubenswrapper[4837]: I1001 08:44:24.207715 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r9jnn" podStartSLOduration=2.769485192 podStartE2EDuration="5.207657967s" podCreationTimestamp="2025-10-01 08:44:19 +0000 UTC" firstStartedPulling="2025-10-01 08:44:21.13322241 +0000 UTC m=+5917.974829905" lastFinishedPulling="2025-10-01 08:44:23.571395225 +0000 UTC m=+5920.413002680" observedRunningTime="2025-10-01 08:44:24.202274934 +0000 UTC m=+5921.043882399" watchObservedRunningTime="2025-10-01 08:44:24.207657967 +0000 UTC m=+5921.049265462" Oct 01 08:44:29 crc kubenswrapper[4837]: I1001 08:44:29.587748 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:29 crc kubenswrapper[4837]: I1001 08:44:29.588372 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:29 crc kubenswrapper[4837]: I1001 08:44:29.652466 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:29 crc kubenswrapper[4837]: I1001 08:44:29.817686 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:44:29 crc kubenswrapper[4837]: E1001 08:44:29.818130 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:44:30 crc kubenswrapper[4837]: I1001 08:44:30.297286 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:30 crc kubenswrapper[4837]: I1001 08:44:30.382034 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.233451 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r9jnn" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="registry-server" containerID="cri-o://a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae" gracePeriod=2 Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.645872 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.843392 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content\") pod \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.843809 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjzjl\" (UniqueName: \"kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl\") pod \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.843940 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities\") pod \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\" (UID: \"b7279ac2-6a40-4342-986b-0fe98eb08f4f\") " Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.844903 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities" (OuterVolumeSpecName: "utilities") pod "b7279ac2-6a40-4342-986b-0fe98eb08f4f" (UID: "b7279ac2-6a40-4342-986b-0fe98eb08f4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.850979 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl" (OuterVolumeSpecName: "kube-api-access-cjzjl") pod "b7279ac2-6a40-4342-986b-0fe98eb08f4f" (UID: "b7279ac2-6a40-4342-986b-0fe98eb08f4f"). InnerVolumeSpecName "kube-api-access-cjzjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.945346 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjzjl\" (UniqueName: \"kubernetes.io/projected/b7279ac2-6a40-4342-986b-0fe98eb08f4f-kube-api-access-cjzjl\") on node \"crc\" DevicePath \"\"" Oct 01 08:44:32 crc kubenswrapper[4837]: I1001 08:44:32.945379 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.244233 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerID="a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae" exitCode=0 Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.244283 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerDied","Data":"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae"} Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.244324 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9jnn" event={"ID":"b7279ac2-6a40-4342-986b-0fe98eb08f4f","Type":"ContainerDied","Data":"9b05457a4528a0f3918c492b52f958d762eb805a40eb71f7c78c428217cb7612"} Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.244324 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9jnn" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.244345 4837 scope.go:117] "RemoveContainer" containerID="a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.272160 4837 scope.go:117] "RemoveContainer" containerID="ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.308041 4837 scope.go:117] "RemoveContainer" containerID="d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.351552 4837 scope.go:117] "RemoveContainer" containerID="a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae" Oct 01 08:44:33 crc kubenswrapper[4837]: E1001 08:44:33.352283 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae\": container with ID starting with a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae not found: ID does not exist" containerID="a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.352348 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae"} err="failed to get container status \"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae\": rpc error: code = NotFound desc = could not find container \"a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae\": container with ID starting with a4b01a3420724039e6542df7735596cd06b9d84a6abbac3b91032ac6a1e798ae not found: ID does not exist" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.352393 4837 scope.go:117] "RemoveContainer" containerID="ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508" Oct 01 08:44:33 crc kubenswrapper[4837]: E1001 08:44:33.352861 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508\": container with ID starting with ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508 not found: ID does not exist" containerID="ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.352933 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508"} err="failed to get container status \"ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508\": rpc error: code = NotFound desc = could not find container \"ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508\": container with ID starting with ac2ad3efba76871179fe8f05216eb5e4f9b180703cdeb018df708c1fcf1c1508 not found: ID does not exist" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.352979 4837 scope.go:117] "RemoveContainer" containerID="d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd" Oct 01 08:44:33 crc kubenswrapper[4837]: E1001 08:44:33.354213 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd\": container with ID starting with d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd not found: ID does not exist" containerID="d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd" Oct 01 08:44:33 crc kubenswrapper[4837]: I1001 08:44:33.354258 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd"} err="failed to get container status \"d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd\": rpc error: code = NotFound desc = could not find container \"d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd\": container with ID starting with d1829a9892ad46dcad7b1887ced8b4de64c395a9c376e50f1b547c28646bf3cd not found: ID does not exist" Oct 01 08:44:34 crc kubenswrapper[4837]: I1001 08:44:34.384935 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7279ac2-6a40-4342-986b-0fe98eb08f4f" (UID: "b7279ac2-6a40-4342-986b-0fe98eb08f4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:44:34 crc kubenswrapper[4837]: I1001 08:44:34.469001 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7279ac2-6a40-4342-986b-0fe98eb08f4f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:44:34 crc kubenswrapper[4837]: I1001 08:44:34.517299 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:34 crc kubenswrapper[4837]: I1001 08:44:34.528899 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r9jnn"] Oct 01 08:44:35 crc kubenswrapper[4837]: I1001 08:44:35.829129 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" path="/var/lib/kubelet/pods/b7279ac2-6a40-4342-986b-0fe98eb08f4f/volumes" Oct 01 08:44:42 crc kubenswrapper[4837]: I1001 08:44:42.816179 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:44:42 crc kubenswrapper[4837]: E1001 08:44:42.817147 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:44:56 crc kubenswrapper[4837]: I1001 08:44:56.816461 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:44:57 crc kubenswrapper[4837]: I1001 08:44:57.480109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada"} Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.164657 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66"] Oct 01 08:45:00 crc kubenswrapper[4837]: E1001 08:45:00.165501 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="registry-server" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.165513 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="registry-server" Oct 01 08:45:00 crc kubenswrapper[4837]: E1001 08:45:00.165530 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="extract-utilities" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.165536 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="extract-utilities" Oct 01 08:45:00 crc kubenswrapper[4837]: E1001 08:45:00.165552 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="extract-content" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.165558 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="extract-content" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.165726 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7279ac2-6a40-4342-986b-0fe98eb08f4f" containerName="registry-server" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.166246 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.169084 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.170153 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.171180 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66"] Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.288103 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2g9\" (UniqueName: \"kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.288870 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.289145 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.390610 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2g9\" (UniqueName: \"kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.391014 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.391059 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.391901 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.404160 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.416010 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2g9\" (UniqueName: \"kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9\") pod \"collect-profiles-29321805-mng66\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.500585 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:00 crc kubenswrapper[4837]: W1001 08:45:00.992329 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30a3802d_4081_433c_a6e1_40f2f0c2f0b0.slice/crio-995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d WatchSource:0}: Error finding container 995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d: Status 404 returned error can't find the container with id 995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d Oct 01 08:45:00 crc kubenswrapper[4837]: I1001 08:45:00.992808 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66"] Oct 01 08:45:01 crc kubenswrapper[4837]: I1001 08:45:01.515106 4837 generic.go:334] "Generic (PLEG): container finished" podID="30a3802d-4081-433c-a6e1-40f2f0c2f0b0" containerID="58890a460f768b7ab805c5c239375dcab797cac9f839f99c4208dc616f49952e" exitCode=0 Oct 01 08:45:01 crc kubenswrapper[4837]: I1001 08:45:01.515167 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" event={"ID":"30a3802d-4081-433c-a6e1-40f2f0c2f0b0","Type":"ContainerDied","Data":"58890a460f768b7ab805c5c239375dcab797cac9f839f99c4208dc616f49952e"} Oct 01 08:45:01 crc kubenswrapper[4837]: I1001 08:45:01.515519 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" event={"ID":"30a3802d-4081-433c-a6e1-40f2f0c2f0b0","Type":"ContainerStarted","Data":"995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d"} Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.791082 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.933983 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs2g9\" (UniqueName: \"kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9\") pod \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.934103 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume\") pod \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.934184 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume\") pod \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\" (UID: \"30a3802d-4081-433c-a6e1-40f2f0c2f0b0\") " Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.935159 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "30a3802d-4081-433c-a6e1-40f2f0c2f0b0" (UID: "30a3802d-4081-433c-a6e1-40f2f0c2f0b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.947248 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "30a3802d-4081-433c-a6e1-40f2f0c2f0b0" (UID: "30a3802d-4081-433c-a6e1-40f2f0c2f0b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:45:02 crc kubenswrapper[4837]: I1001 08:45:02.947262 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9" (OuterVolumeSpecName: "kube-api-access-bs2g9") pod "30a3802d-4081-433c-a6e1-40f2f0c2f0b0" (UID: "30a3802d-4081-433c-a6e1-40f2f0c2f0b0"). InnerVolumeSpecName "kube-api-access-bs2g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.036778 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs2g9\" (UniqueName: \"kubernetes.io/projected/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-kube-api-access-bs2g9\") on node \"crc\" DevicePath \"\"" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.037037 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.037148 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/30a3802d-4081-433c-a6e1-40f2f0c2f0b0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.533335 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" event={"ID":"30a3802d-4081-433c-a6e1-40f2f0c2f0b0","Type":"ContainerDied","Data":"995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d"} Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.533382 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="995c57c6749c0ad97aae3915e99d96726c7ef2cc41f197c895b57789d802e65d" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.533448 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66" Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.873972 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d"] Oct 01 08:45:03 crc kubenswrapper[4837]: I1001 08:45:03.878941 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321760-x524d"] Oct 01 08:45:05 crc kubenswrapper[4837]: I1001 08:45:05.836489 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdc43b23-0756-47b2-94e8-267b8746967d" path="/var/lib/kubelet/pods/fdc43b23-0756-47b2-94e8-267b8746967d/volumes" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.191017 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:02 crc kubenswrapper[4837]: E1001 08:46:02.192385 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a3802d-4081-433c-a6e1-40f2f0c2f0b0" containerName="collect-profiles" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.192407 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a3802d-4081-433c-a6e1-40f2f0c2f0b0" containerName="collect-profiles" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.192731 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a3802d-4081-433c-a6e1-40f2f0c2f0b0" containerName="collect-profiles" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.198278 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.209556 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.223126 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.223223 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s5xr\" (UniqueName: \"kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.223259 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.324433 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.324517 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s5xr\" (UniqueName: \"kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.324541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.325062 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.325424 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.349448 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s5xr\" (UniqueName: \"kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr\") pod \"redhat-marketplace-c8qzm\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.541658 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:02 crc kubenswrapper[4837]: I1001 08:46:02.826255 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:03 crc kubenswrapper[4837]: I1001 08:46:03.180163 4837 generic.go:334] "Generic (PLEG): container finished" podID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerID="f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9" exitCode=0 Oct 01 08:46:03 crc kubenswrapper[4837]: I1001 08:46:03.180210 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerDied","Data":"f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9"} Oct 01 08:46:03 crc kubenswrapper[4837]: I1001 08:46:03.180234 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerStarted","Data":"fa478424263f6f6a77de9c5196e5c727c914bd301f481845a36c1989698c6566"} Oct 01 08:46:04 crc kubenswrapper[4837]: I1001 08:46:04.188740 4837 generic.go:334] "Generic (PLEG): container finished" podID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerID="f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5" exitCode=0 Oct 01 08:46:04 crc kubenswrapper[4837]: I1001 08:46:04.188840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerDied","Data":"f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5"} Oct 01 08:46:05 crc kubenswrapper[4837]: I1001 08:46:05.204854 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerStarted","Data":"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864"} Oct 01 08:46:05 crc kubenswrapper[4837]: I1001 08:46:05.437569 4837 scope.go:117] "RemoveContainer" containerID="807e99129e0fd1f1d04280f79b08d2b4793b3c4fc4db205f6f81799243bfd259" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.004586 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c8qzm" podStartSLOduration=7.605326514 podStartE2EDuration="9.004558255s" podCreationTimestamp="2025-10-01 08:46:02 +0000 UTC" firstStartedPulling="2025-10-01 08:46:03.182349693 +0000 UTC m=+6020.023957148" lastFinishedPulling="2025-10-01 08:46:04.581581424 +0000 UTC m=+6021.423188889" observedRunningTime="2025-10-01 08:46:05.230093518 +0000 UTC m=+6022.071701013" watchObservedRunningTime="2025-10-01 08:46:11.004558255 +0000 UTC m=+6027.846165710" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.009940 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.011561 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.017963 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.019289 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.021368 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-sdv54" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.021531 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.021560 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.021543 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.021578 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.032763 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.042980 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.173441 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.173496 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8sxm\" (UniqueName: \"kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.173534 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.173577 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g45sr\" (UniqueName: \"kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.173849 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.275988 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.276039 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8sxm\" (UniqueName: \"kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.276075 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.276111 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g45sr\" (UniqueName: \"kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.276149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.277341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.277341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.277355 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.305229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g45sr\" (UniqueName: \"kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr\") pod \"dnsmasq-dns-85c9546d75-dhldj\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.316135 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8sxm\" (UniqueName: \"kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm\") pod \"dnsmasq-dns-6499b68dfc-7xnfl\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.325529 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.326273 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.353608 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.354427 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.361608 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.373574 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.483528 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.484258 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.484793 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7pst\" (UniqueName: \"kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.587613 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.587713 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7pst\" (UniqueName: \"kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.587777 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.589328 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.594297 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.625254 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7pst\" (UniqueName: \"kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst\") pod \"dnsmasq-dns-6d98fd957f-h677b\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.693902 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.728367 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.731987 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.753604 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.774346 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.893100 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.893170 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.893519 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srzj4\" (UniqueName: \"kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.957039 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.995356 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srzj4\" (UniqueName: \"kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.995475 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.995521 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.998575 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:11 crc kubenswrapper[4837]: I1001 08:46:11.999460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.024524 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srzj4\" (UniqueName: \"kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4\") pod \"dnsmasq-dns-689c49c455-7kv79\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.043660 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:12 crc kubenswrapper[4837]: W1001 08:46:12.047998 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c33f13b_00b7_48ae_9151_b77dfe90ad21.slice/crio-af9e42070ea4b0685e84a85bd1c59f61ceaf0f91e84b2c7514977b4a8ff483c7 WatchSource:0}: Error finding container af9e42070ea4b0685e84a85bd1c59f61ceaf0f91e84b2c7514977b4a8ff483c7: Status 404 returned error can't find the container with id af9e42070ea4b0685e84a85bd1c59f61ceaf0f91e84b2c7514977b4a8ff483c7 Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.064743 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.279443 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" event={"ID":"0c33f13b-00b7-48ae-9151-b77dfe90ad21","Type":"ContainerStarted","Data":"af9e42070ea4b0685e84a85bd1c59f61ceaf0f91e84b2c7514977b4a8ff483c7"} Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.281818 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" event={"ID":"c1083379-0b1d-4003-81e4-6cc8d9839478","Type":"ContainerStarted","Data":"003eccd98b9cdd3743d9316e968c0c8bbd3e8a180141543c7b85334bcb80c69f"} Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.340376 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.524899 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.526753 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.537370 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.537756 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.538030 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.538156 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.538213 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.539027 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ts9d5" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.538618 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.542312 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.543474 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.582019 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.638446 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.644316 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650674 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650742 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650806 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhkfx\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650830 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650846 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650883 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650926 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650959 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.650981 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.651001 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.651050 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752461 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752529 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752581 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752639 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752677 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752758 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhkfx\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752801 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752816 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752831 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752847 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.752884 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.754677 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.754923 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.755872 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.757156 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.757832 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.767820 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.772552 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.773931 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.773988 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afbb63deeecbf25cf80875790908ff7c42a9c13794df34babbfb4f9f2a1e87b2/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.776016 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhkfx\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.780537 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.781244 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.845388 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.870865 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.873504 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.878530 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.882797 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.882827 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-htgzk" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.883392 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.883414 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.883593 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.883673 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.883677 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 08:46:12 crc kubenswrapper[4837]: I1001 08:46:12.888636 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058454 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058532 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058590 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058616 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058636 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058669 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058740 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058767 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gxvv\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058812 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.058868 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160621 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160714 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160738 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160764 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160806 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.160832 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gxvv\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.161158 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.161338 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.161403 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.161427 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.161457 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.162874 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.163937 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.166126 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.166183 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ad0c11f22fa60721e63eac5a1d2193d37d85e125324ebb604a4e9e68815f42a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.166308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.167598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.167866 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.168152 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.168871 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.169976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.170546 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.179246 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gxvv\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.196581 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.214366 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.334565 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" event={"ID":"eda09a1e-e8ee-4af5-99b3-54601be231d1","Type":"ContainerStarted","Data":"409e63f492aee87350779c5448a8739d62db42039733ee6131ee601fa277bd90"} Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.338332 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.340382 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689c49c455-7kv79" event={"ID":"0be6f3e3-91de-4d71-abe7-d1b2beed5476","Type":"ContainerStarted","Data":"e0a67a0abe38859944070c94fadd32a984a466b1894a4dfcb57bc29e5ee4bb6d"} Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.471560 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.535923 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.751631 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:46:13 crc kubenswrapper[4837]: W1001 08:46:13.775976 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc216a3dd_cce4_4c8a_89c4_7f65c1ba63f7.slice/crio-925cefbc3ed7cb0c13912c708203da969027054d9502461f422c128745da3336 WatchSource:0}: Error finding container 925cefbc3ed7cb0c13912c708203da969027054d9502461f422c128745da3336: Status 404 returned error can't find the container with id 925cefbc3ed7cb0c13912c708203da969027054d9502461f422c128745da3336 Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.809363 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.810930 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.816341 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.823884 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-smstf" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.826310 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.826630 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.826803 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.827783 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.877363 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975257 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-default\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975361 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975404 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975459 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-secrets\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975499 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975534 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975572 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975597 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xnzs\" (UniqueName: \"kubernetes.io/projected/5bced997-a865-4e08-b41d-53f9ed5ea466-kube-api-access-6xnzs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:13 crc kubenswrapper[4837]: I1001 08:46:13.975627 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-kolla-config\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077253 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077340 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-secrets\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077366 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077392 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077411 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077434 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xnzs\" (UniqueName: \"kubernetes.io/projected/5bced997-a865-4e08-b41d-53f9ed5ea466-kube-api-access-6xnzs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077451 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-kolla-config\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.077498 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-default\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.078179 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.078978 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.081317 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-kolla-config\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.083180 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.084231 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.084299 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0f2965769b94a7ec43b78746dac0c841b2e0560a5d8785fb818031873612bbad/globalmount\"" pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.091430 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5bced997-a865-4e08-b41d-53f9ed5ea466-config-data-default\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.096055 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-secrets\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.101635 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xnzs\" (UniqueName: \"kubernetes.io/projected/5bced997-a865-4e08-b41d-53f9ed5ea466-kube-api-access-6xnzs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.107385 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bced997-a865-4e08-b41d-53f9ed5ea466-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.125428 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-517781ab-e8df-46a2-bb25-070f3d90fd14\") pod \"openstack-galera-0\" (UID: \"5bced997-a865-4e08-b41d-53f9ed5ea466\") " pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.162073 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.465764 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerStarted","Data":"925cefbc3ed7cb0c13912c708203da969027054d9502461f422c128745da3336"} Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.481042 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerStarted","Data":"c8c05fed940c5ef91b66e412eae28328a6a3f2365f5fb1aeda6df34842413e36"} Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.757922 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 08:46:14 crc kubenswrapper[4837]: W1001 08:46:14.770847 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bced997_a865_4e08_b41d_53f9ed5ea466.slice/crio-4b7433da111064816e11540c17e3b62b7e65c86c94b43e3deb645f8d1ef8e83f WatchSource:0}: Error finding container 4b7433da111064816e11540c17e3b62b7e65c86c94b43e3deb645f8d1ef8e83f: Status 404 returned error can't find the container with id 4b7433da111064816e11540c17e3b62b7e65c86c94b43e3deb645f8d1ef8e83f Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.853886 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.857960 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.861233 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.861737 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.861954 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9kxd8" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.862226 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 01 08:46:14 crc kubenswrapper[4837]: I1001 08:46:14.871008 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001379 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001433 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9tdt\" (UniqueName: \"kubernetes.io/projected/001d3974-ad2f-4ef1-880c-ce4e32600f24-kube-api-access-f9tdt\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001487 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001540 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001574 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001631 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001665 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.001704 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.103976 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104039 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104088 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104118 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104144 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104238 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104298 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9tdt\" (UniqueName: \"kubernetes.io/projected/001d3974-ad2f-4ef1-880c-ce4e32600f24-kube-api-access-f9tdt\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.104333 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.105591 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.105966 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.106044 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.106554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/001d3974-ad2f-4ef1-880c-ce4e32600f24-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.109151 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.109215 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2f58ca5d72db8a76a36055b99d2ee1c27b1df30a2d7bc3241f6499d1933ae959/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.111946 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.117467 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.123496 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001d3974-ad2f-4ef1-880c-ce4e32600f24-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.133358 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9tdt\" (UniqueName: \"kubernetes.io/projected/001d3974-ad2f-4ef1-880c-ce4e32600f24-kube-api-access-f9tdt\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.163516 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-03553b8f-e7a5-484b-94a0-80d135f25bab\") pod \"openstack-cell1-galera-0\" (UID: \"001d3974-ad2f-4ef1-880c-ce4e32600f24\") " pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.188675 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.370498 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.377078 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.380245 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tth64" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.382099 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.385060 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.385326 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.506109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5bced997-a865-4e08-b41d-53f9ed5ea466","Type":"ContainerStarted","Data":"4b7433da111064816e11540c17e3b62b7e65c86c94b43e3deb645f8d1ef8e83f"} Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.506476 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c8qzm" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="registry-server" containerID="cri-o://681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864" gracePeriod=2 Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.511383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.511478 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brzj5\" (UniqueName: \"kubernetes.io/projected/e98b8d09-2148-42d2-aeef-650cc3b41feb-kube-api-access-brzj5\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.511605 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-config-data\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.511636 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.511669 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-kolla-config\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.613605 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-config-data\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.613675 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.613747 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-kolla-config\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.613803 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.613902 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brzj5\" (UniqueName: \"kubernetes.io/projected/e98b8d09-2148-42d2-aeef-650cc3b41feb-kube-api-access-brzj5\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.615229 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-config-data\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.615358 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e98b8d09-2148-42d2-aeef-650cc3b41feb-kolla-config\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.621866 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.624220 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b8d09-2148-42d2-aeef-650cc3b41feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.630469 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brzj5\" (UniqueName: \"kubernetes.io/projected/e98b8d09-2148-42d2-aeef-650cc3b41feb-kube-api-access-brzj5\") pod \"memcached-0\" (UID: \"e98b8d09-2148-42d2-aeef-650cc3b41feb\") " pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.711686 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 08:46:15 crc kubenswrapper[4837]: I1001 08:46:15.759234 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 08:46:15 crc kubenswrapper[4837]: W1001 08:46:15.791444 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod001d3974_ad2f_4ef1_880c_ce4e32600f24.slice/crio-294f3612c5d99def8f70a721e4ae0e50affc6d172693ffffcf95fdf442e99413 WatchSource:0}: Error finding container 294f3612c5d99def8f70a721e4ae0e50affc6d172693ffffcf95fdf442e99413: Status 404 returned error can't find the container with id 294f3612c5d99def8f70a721e4ae0e50affc6d172693ffffcf95fdf442e99413 Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.034833 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.130158 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s5xr\" (UniqueName: \"kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr\") pod \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.130321 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content\") pod \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.131765 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities\") pod \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\" (UID: \"b59f1c24-1509-415c-a312-c4d5e9b02f6c\") " Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.133610 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities" (OuterVolumeSpecName: "utilities") pod "b59f1c24-1509-415c-a312-c4d5e9b02f6c" (UID: "b59f1c24-1509-415c-a312-c4d5e9b02f6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.137633 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr" (OuterVolumeSpecName: "kube-api-access-9s5xr") pod "b59f1c24-1509-415c-a312-c4d5e9b02f6c" (UID: "b59f1c24-1509-415c-a312-c4d5e9b02f6c"). InnerVolumeSpecName "kube-api-access-9s5xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.144301 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b59f1c24-1509-415c-a312-c4d5e9b02f6c" (UID: "b59f1c24-1509-415c-a312-c4d5e9b02f6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.234312 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.234373 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s5xr\" (UniqueName: \"kubernetes.io/projected/b59f1c24-1509-415c-a312-c4d5e9b02f6c-kube-api-access-9s5xr\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.234386 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b59f1c24-1509-415c-a312-c4d5e9b02f6c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.368310 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 08:46:16 crc kubenswrapper[4837]: W1001 08:46:16.388849 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode98b8d09_2148_42d2_aeef_650cc3b41feb.slice/crio-16851c0ec109380a626cb2b87d987b4f428872d992ca514b6fbc134fc236311c WatchSource:0}: Error finding container 16851c0ec109380a626cb2b87d987b4f428872d992ca514b6fbc134fc236311c: Status 404 returned error can't find the container with id 16851c0ec109380a626cb2b87d987b4f428872d992ca514b6fbc134fc236311c Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.519661 4837 generic.go:334] "Generic (PLEG): container finished" podID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerID="681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864" exitCode=0 Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.519786 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerDied","Data":"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864"} Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.519830 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8qzm" event={"ID":"b59f1c24-1509-415c-a312-c4d5e9b02f6c","Type":"ContainerDied","Data":"fa478424263f6f6a77de9c5196e5c727c914bd301f481845a36c1989698c6566"} Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.519857 4837 scope.go:117] "RemoveContainer" containerID="681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.520046 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8qzm" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.529035 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e98b8d09-2148-42d2-aeef-650cc3b41feb","Type":"ContainerStarted","Data":"16851c0ec109380a626cb2b87d987b4f428872d992ca514b6fbc134fc236311c"} Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.531530 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"001d3974-ad2f-4ef1-880c-ce4e32600f24","Type":"ContainerStarted","Data":"294f3612c5d99def8f70a721e4ae0e50affc6d172693ffffcf95fdf442e99413"} Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.560582 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.567846 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8qzm"] Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.573160 4837 scope.go:117] "RemoveContainer" containerID="f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.605520 4837 scope.go:117] "RemoveContainer" containerID="f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.654271 4837 scope.go:117] "RemoveContainer" containerID="681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864" Oct 01 08:46:16 crc kubenswrapper[4837]: E1001 08:46:16.654909 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864\": container with ID starting with 681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864 not found: ID does not exist" containerID="681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.654943 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864"} err="failed to get container status \"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864\": rpc error: code = NotFound desc = could not find container \"681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864\": container with ID starting with 681836e96ffa05ce82805e3cb65e0aeb4a5b00d3e67059bf380dd21d72150864 not found: ID does not exist" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.654966 4837 scope.go:117] "RemoveContainer" containerID="f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5" Oct 01 08:46:16 crc kubenswrapper[4837]: E1001 08:46:16.655350 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5\": container with ID starting with f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5 not found: ID does not exist" containerID="f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.655378 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5"} err="failed to get container status \"f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5\": rpc error: code = NotFound desc = could not find container \"f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5\": container with ID starting with f2084f10ca5f6510194ff81a6d03c58a43bd626aed130dcda7e985fff6ba37f5 not found: ID does not exist" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.655394 4837 scope.go:117] "RemoveContainer" containerID="f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9" Oct 01 08:46:16 crc kubenswrapper[4837]: E1001 08:46:16.655746 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9\": container with ID starting with f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9 not found: ID does not exist" containerID="f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9" Oct 01 08:46:16 crc kubenswrapper[4837]: I1001 08:46:16.655768 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9"} err="failed to get container status \"f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9\": rpc error: code = NotFound desc = could not find container \"f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9\": container with ID starting with f282bea1a82fafbeea8ccb06b8179df9b11973da78374bc55b38851923b18bb9 not found: ID does not exist" Oct 01 08:46:17 crc kubenswrapper[4837]: I1001 08:46:17.826596 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" path="/var/lib/kubelet/pods/b59f1c24-1509-415c-a312-c4d5e9b02f6c/volumes" Oct 01 08:46:38 crc kubenswrapper[4837]: E1001 08:46:38.459160 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:38 crc kubenswrapper[4837]: E1001 08:46:38.459817 4837 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:38 crc kubenswrapper[4837]: E1001 08:46:38.460000 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l7pst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6d98fd957f-h677b_openstack(eda09a1e-e8ee-4af5-99b3-54601be231d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 08:46:38 crc kubenswrapper[4837]: E1001 08:46:38.461911 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" Oct 01 08:46:38 crc kubenswrapper[4837]: E1001 08:46:38.750640 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a\\\"\"" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" Oct 01 08:46:40 crc kubenswrapper[4837]: E1001 08:46:40.410654 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:40 crc kubenswrapper[4837]: E1001 08:46:40.410766 4837 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:40 crc kubenswrapper[4837]: E1001 08:46:40.410951 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n564h564h676h699hcdh67bh66hfdh569h545h648h94h546h696h668h89h96h667h575h595h5d9h584h8dhbdh697h54bhb7h58fh5c9hd8h5cdh5c7q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-srzj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-689c49c455-7kv79_openstack(0be6f3e3-91de-4d71-abe7-d1b2beed5476): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 08:46:40 crc kubenswrapper[4837]: E1001 08:46:40.412245 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-689c49c455-7kv79" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" Oct 01 08:46:40 crc kubenswrapper[4837]: E1001 08:46:40.770401 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:a0eac564d779a7eaac46c9816bff261a\\\"\"" pod="openstack/dnsmasq-dns-689c49c455-7kv79" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.433560 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.433711 4837 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.433937 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gxvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.435303 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.451645 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.451746 4837 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.452001 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhkfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(b86b0277-4772-4869-93dc-e261cd56b06f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.453319 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" Oct 01 08:46:41 crc kubenswrapper[4837]: I1001 08:46:41.807648 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"001d3974-ad2f-4ef1-880c-ce4e32600f24","Type":"ContainerStarted","Data":"2ab49c4d853c1a2a24d1ea7dcfe1e67c90c9df072654ae78164395e85da04eaf"} Oct 01 08:46:41 crc kubenswrapper[4837]: I1001 08:46:41.813112 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 01 08:46:41 crc kubenswrapper[4837]: I1001 08:46:41.816221 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" podUID="c1083379-0b1d-4003-81e4-6cc8d9839478" containerName="init" containerID="cri-o://32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b" gracePeriod=10 Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.822546 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a\\\"\"" pod="openstack/rabbitmq-server-0" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" Oct 01 08:46:41 crc kubenswrapper[4837]: E1001 08:46:41.822676 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:a0eac564d779a7eaac46c9816bff261a\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" Oct 01 08:46:41 crc kubenswrapper[4837]: I1001 08:46:41.829616 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" event={"ID":"0c33f13b-00b7-48ae-9151-b77dfe90ad21","Type":"ContainerStarted","Data":"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18"} Oct 01 08:46:41 crc kubenswrapper[4837]: I1001 08:46:41.903399 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.8397089150000001 podStartE2EDuration="26.903373906s" podCreationTimestamp="2025-10-01 08:46:15 +0000 UTC" firstStartedPulling="2025-10-01 08:46:16.392757127 +0000 UTC m=+6033.234364582" lastFinishedPulling="2025-10-01 08:46:41.456422118 +0000 UTC m=+6058.298029573" observedRunningTime="2025-10-01 08:46:41.893745459 +0000 UTC m=+6058.735352904" watchObservedRunningTime="2025-10-01 08:46:41.903373906 +0000 UTC m=+6058.744981361" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.155235 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.212942 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.266307 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config\") pod \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.266416 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config\") pod \"c1083379-0b1d-4003-81e4-6cc8d9839478\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.266474 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc\") pod \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.266711 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8sxm\" (UniqueName: \"kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm\") pod \"c1083379-0b1d-4003-81e4-6cc8d9839478\" (UID: \"c1083379-0b1d-4003-81e4-6cc8d9839478\") " Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.266809 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g45sr\" (UniqueName: \"kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr\") pod \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\" (UID: \"0c33f13b-00b7-48ae-9151-b77dfe90ad21\") " Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.272342 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr" (OuterVolumeSpecName: "kube-api-access-g45sr") pod "0c33f13b-00b7-48ae-9151-b77dfe90ad21" (UID: "0c33f13b-00b7-48ae-9151-b77dfe90ad21"). InnerVolumeSpecName "kube-api-access-g45sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.272458 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm" (OuterVolumeSpecName: "kube-api-access-g8sxm") pod "c1083379-0b1d-4003-81e4-6cc8d9839478" (UID: "c1083379-0b1d-4003-81e4-6cc8d9839478"). InnerVolumeSpecName "kube-api-access-g8sxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.288090 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config" (OuterVolumeSpecName: "config") pod "0c33f13b-00b7-48ae-9151-b77dfe90ad21" (UID: "0c33f13b-00b7-48ae-9151-b77dfe90ad21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.289809 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c33f13b-00b7-48ae-9151-b77dfe90ad21" (UID: "0c33f13b-00b7-48ae-9151-b77dfe90ad21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.290563 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config" (OuterVolumeSpecName: "config") pod "c1083379-0b1d-4003-81e4-6cc8d9839478" (UID: "c1083379-0b1d-4003-81e4-6cc8d9839478"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.370485 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8sxm\" (UniqueName: \"kubernetes.io/projected/c1083379-0b1d-4003-81e4-6cc8d9839478-kube-api-access-g8sxm\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.370879 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g45sr\" (UniqueName: \"kubernetes.io/projected/0c33f13b-00b7-48ae-9151-b77dfe90ad21-kube-api-access-g45sr\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.370944 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.370998 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1083379-0b1d-4003-81e4-6cc8d9839478-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.371048 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33f13b-00b7-48ae-9151-b77dfe90ad21-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.829577 4837 generic.go:334] "Generic (PLEG): container finished" podID="0c33f13b-00b7-48ae-9151-b77dfe90ad21" containerID="4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18" exitCode=0 Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.829673 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.829686 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" event={"ID":"0c33f13b-00b7-48ae-9151-b77dfe90ad21","Type":"ContainerDied","Data":"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.830353 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c9546d75-dhldj" event={"ID":"0c33f13b-00b7-48ae-9151-b77dfe90ad21","Type":"ContainerDied","Data":"af9e42070ea4b0685e84a85bd1c59f61ceaf0f91e84b2c7514977b4a8ff483c7"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.830491 4837 scope.go:117] "RemoveContainer" containerID="4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.834371 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5bced997-a865-4e08-b41d-53f9ed5ea466","Type":"ContainerStarted","Data":"d617927785a7275b0a12a10fab9bb259e4345397e249bb0b7ca6d5e6864a95f3"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.838413 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e98b8d09-2148-42d2-aeef-650cc3b41feb","Type":"ContainerStarted","Data":"cf39888b80dcc7dd4f9b8b4e78f0fd701388c14364e5df4b9ff5f13e7e97e534"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.841674 4837 generic.go:334] "Generic (PLEG): container finished" podID="c1083379-0b1d-4003-81e4-6cc8d9839478" containerID="32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b" exitCode=0 Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.841844 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.841916 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" event={"ID":"c1083379-0b1d-4003-81e4-6cc8d9839478","Type":"ContainerDied","Data":"32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.841961 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6499b68dfc-7xnfl" event={"ID":"c1083379-0b1d-4003-81e4-6cc8d9839478","Type":"ContainerDied","Data":"003eccd98b9cdd3743d9316e968c0c8bbd3e8a180141543c7b85334bcb80c69f"} Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.859799 4837 scope.go:117] "RemoveContainer" containerID="4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18" Oct 01 08:46:42 crc kubenswrapper[4837]: E1001 08:46:42.865130 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18\": container with ID starting with 4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18 not found: ID does not exist" containerID="4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.865187 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18"} err="failed to get container status \"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18\": rpc error: code = NotFound desc = could not find container \"4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18\": container with ID starting with 4c70754e7127f4a1595d7e7e2192ae957fc13505f969331aa3f1ca811ab44e18 not found: ID does not exist" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.865223 4837 scope.go:117] "RemoveContainer" containerID="32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.903296 4837 scope.go:117] "RemoveContainer" containerID="32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b" Oct 01 08:46:42 crc kubenswrapper[4837]: E1001 08:46:42.909319 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b\": container with ID starting with 32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b not found: ID does not exist" containerID="32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.909427 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b"} err="failed to get container status \"32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b\": rpc error: code = NotFound desc = could not find container \"32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b\": container with ID starting with 32f695fdbb04b72dc39f9395551af57ae37c5207f3c51a3aed505c0e30bdc89b not found: ID does not exist" Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.935895 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.945142 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85c9546d75-dhldj"] Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.982072 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:42 crc kubenswrapper[4837]: I1001 08:46:42.989496 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6499b68dfc-7xnfl"] Oct 01 08:46:43 crc kubenswrapper[4837]: I1001 08:46:43.845767 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c33f13b-00b7-48ae-9151-b77dfe90ad21" path="/var/lib/kubelet/pods/0c33f13b-00b7-48ae-9151-b77dfe90ad21/volumes" Oct 01 08:46:43 crc kubenswrapper[4837]: I1001 08:46:43.847315 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1083379-0b1d-4003-81e4-6cc8d9839478" path="/var/lib/kubelet/pods/c1083379-0b1d-4003-81e4-6cc8d9839478/volumes" Oct 01 08:46:45 crc kubenswrapper[4837]: I1001 08:46:45.880938 4837 generic.go:334] "Generic (PLEG): container finished" podID="001d3974-ad2f-4ef1-880c-ce4e32600f24" containerID="2ab49c4d853c1a2a24d1ea7dcfe1e67c90c9df072654ae78164395e85da04eaf" exitCode=0 Oct 01 08:46:45 crc kubenswrapper[4837]: I1001 08:46:45.881467 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"001d3974-ad2f-4ef1-880c-ce4e32600f24","Type":"ContainerDied","Data":"2ab49c4d853c1a2a24d1ea7dcfe1e67c90c9df072654ae78164395e85da04eaf"} Oct 01 08:46:45 crc kubenswrapper[4837]: I1001 08:46:45.884971 4837 generic.go:334] "Generic (PLEG): container finished" podID="5bced997-a865-4e08-b41d-53f9ed5ea466" containerID="d617927785a7275b0a12a10fab9bb259e4345397e249bb0b7ca6d5e6864a95f3" exitCode=0 Oct 01 08:46:45 crc kubenswrapper[4837]: I1001 08:46:45.885010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5bced997-a865-4e08-b41d-53f9ed5ea466","Type":"ContainerDied","Data":"d617927785a7275b0a12a10fab9bb259e4345397e249bb0b7ca6d5e6864a95f3"} Oct 01 08:46:46 crc kubenswrapper[4837]: I1001 08:46:46.911107 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"001d3974-ad2f-4ef1-880c-ce4e32600f24","Type":"ContainerStarted","Data":"3b7a288896880cdc2a23b5b659beda9e2d4dc5ac7aab8f01d2b8aeac353d22d3"} Oct 01 08:46:46 crc kubenswrapper[4837]: I1001 08:46:46.914320 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5bced997-a865-4e08-b41d-53f9ed5ea466","Type":"ContainerStarted","Data":"bbc5d76004a43c4da31e9cb896ffd2a3405b3e9915f7fb7255cff196c3bdba85"} Oct 01 08:46:46 crc kubenswrapper[4837]: I1001 08:46:46.946187 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.203555823 podStartE2EDuration="33.946153515s" podCreationTimestamp="2025-10-01 08:46:13 +0000 UTC" firstStartedPulling="2025-10-01 08:46:15.796699263 +0000 UTC m=+6032.638306718" lastFinishedPulling="2025-10-01 08:46:41.539296955 +0000 UTC m=+6058.380904410" observedRunningTime="2025-10-01 08:46:46.942901565 +0000 UTC m=+6063.784509050" watchObservedRunningTime="2025-10-01 08:46:46.946153515 +0000 UTC m=+6063.787760980" Oct 01 08:46:46 crc kubenswrapper[4837]: I1001 08:46:46.973214 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.279582915 podStartE2EDuration="34.973192299s" podCreationTimestamp="2025-10-01 08:46:12 +0000 UTC" firstStartedPulling="2025-10-01 08:46:14.777765232 +0000 UTC m=+6031.619372687" lastFinishedPulling="2025-10-01 08:46:41.471374596 +0000 UTC m=+6058.312982071" observedRunningTime="2025-10-01 08:46:46.969298413 +0000 UTC m=+6063.810905958" watchObservedRunningTime="2025-10-01 08:46:46.973192299 +0000 UTC m=+6063.814799764" Oct 01 08:46:50 crc kubenswrapper[4837]: I1001 08:46:50.713470 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 01 08:46:52 crc kubenswrapper[4837]: I1001 08:46:52.818737 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:46:52 crc kubenswrapper[4837]: I1001 08:46:52.972472 4837 generic.go:334] "Generic (PLEG): container finished" podID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerID="c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5" exitCode=0 Oct 01 08:46:52 crc kubenswrapper[4837]: I1001 08:46:52.972559 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" event={"ID":"eda09a1e-e8ee-4af5-99b3-54601be231d1","Type":"ContainerDied","Data":"c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5"} Oct 01 08:46:52 crc kubenswrapper[4837]: I1001 08:46:52.977245 4837 generic.go:334] "Generic (PLEG): container finished" podID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerID="f1350c91350dba84024084a37f54eff5fa68ddfeabb695465b639d4fc3036fbd" exitCode=0 Oct 01 08:46:52 crc kubenswrapper[4837]: I1001 08:46:52.977325 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689c49c455-7kv79" event={"ID":"0be6f3e3-91de-4d71-abe7-d1b2beed5476","Type":"ContainerDied","Data":"f1350c91350dba84024084a37f54eff5fa68ddfeabb695465b639d4fc3036fbd"} Oct 01 08:46:54 crc kubenswrapper[4837]: I1001 08:46:54.163285 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 01 08:46:54 crc kubenswrapper[4837]: I1001 08:46:54.164903 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 01 08:46:54 crc kubenswrapper[4837]: I1001 08:46:54.226851 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 01 08:46:55 crc kubenswrapper[4837]: I1001 08:46:55.052040 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 01 08:46:55 crc kubenswrapper[4837]: I1001 08:46:55.189772 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:55 crc kubenswrapper[4837]: I1001 08:46:55.191996 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:55 crc kubenswrapper[4837]: I1001 08:46:55.244939 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:56 crc kubenswrapper[4837]: I1001 08:46:56.065979 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 01 08:46:57 crc kubenswrapper[4837]: I1001 08:46:57.027403 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" event={"ID":"eda09a1e-e8ee-4af5-99b3-54601be231d1","Type":"ContainerStarted","Data":"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1"} Oct 01 08:46:58 crc kubenswrapper[4837]: I1001 08:46:58.038175 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689c49c455-7kv79" event={"ID":"0be6f3e3-91de-4d71-abe7-d1b2beed5476","Type":"ContainerStarted","Data":"c7140bb7a671a4a474642202d44fd4b2fa5c4a4818b629fe055a94e36fc9a360"} Oct 01 08:46:58 crc kubenswrapper[4837]: I1001 08:46:58.038668 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:46:58 crc kubenswrapper[4837]: I1001 08:46:58.070321 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" podStartSLOduration=-9223371989.784487 podStartE2EDuration="47.070289124s" podCreationTimestamp="2025-10-01 08:46:11 +0000 UTC" firstStartedPulling="2025-10-01 08:46:12.371053371 +0000 UTC m=+6029.212660826" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:46:58.067371223 +0000 UTC m=+6074.908978698" watchObservedRunningTime="2025-10-01 08:46:58.070289124 +0000 UTC m=+6074.911896579" Oct 01 08:46:58 crc kubenswrapper[4837]: I1001 08:46:58.094638 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-689c49c455-7kv79" podStartSLOduration=-9223371989.76017 podStartE2EDuration="47.094606682s" podCreationTimestamp="2025-10-01 08:46:11 +0000 UTC" firstStartedPulling="2025-10-01 08:46:12.654634012 +0000 UTC m=+6029.496241467" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:46:58.087011315 +0000 UTC m=+6074.928618800" watchObservedRunningTime="2025-10-01 08:46:58.094606682 +0000 UTC m=+6074.936214137" Oct 01 08:46:59 crc kubenswrapper[4837]: I1001 08:46:59.054099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerStarted","Data":"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5"} Oct 01 08:46:59 crc kubenswrapper[4837]: I1001 08:46:59.058822 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerStarted","Data":"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe"} Oct 01 08:47:01 crc kubenswrapper[4837]: I1001 08:47:01.777043 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.066167 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.067256 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.159856 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.160159 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="dnsmasq-dns" containerID="cri-o://f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1" gracePeriod=10 Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.584245 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.621705 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config\") pod \"eda09a1e-e8ee-4af5-99b3-54601be231d1\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.621774 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc\") pod \"eda09a1e-e8ee-4af5-99b3-54601be231d1\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.621801 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7pst\" (UniqueName: \"kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst\") pod \"eda09a1e-e8ee-4af5-99b3-54601be231d1\" (UID: \"eda09a1e-e8ee-4af5-99b3-54601be231d1\") " Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.628449 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst" (OuterVolumeSpecName: "kube-api-access-l7pst") pod "eda09a1e-e8ee-4af5-99b3-54601be231d1" (UID: "eda09a1e-e8ee-4af5-99b3-54601be231d1"). InnerVolumeSpecName "kube-api-access-l7pst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.668655 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config" (OuterVolumeSpecName: "config") pod "eda09a1e-e8ee-4af5-99b3-54601be231d1" (UID: "eda09a1e-e8ee-4af5-99b3-54601be231d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.670022 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eda09a1e-e8ee-4af5-99b3-54601be231d1" (UID: "eda09a1e-e8ee-4af5-99b3-54601be231d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.723680 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.723735 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eda09a1e-e8ee-4af5-99b3-54601be231d1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:47:02 crc kubenswrapper[4837]: I1001 08:47:02.723748 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7pst\" (UniqueName: \"kubernetes.io/projected/eda09a1e-e8ee-4af5-99b3-54601be231d1-kube-api-access-l7pst\") on node \"crc\" DevicePath \"\"" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.106350 4837 generic.go:334] "Generic (PLEG): container finished" podID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerID="f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1" exitCode=0 Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.106429 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" event={"ID":"eda09a1e-e8ee-4af5-99b3-54601be231d1","Type":"ContainerDied","Data":"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1"} Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.106903 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" event={"ID":"eda09a1e-e8ee-4af5-99b3-54601be231d1","Type":"ContainerDied","Data":"409e63f492aee87350779c5448a8739d62db42039733ee6131ee601fa277bd90"} Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.106477 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d98fd957f-h677b" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.106945 4837 scope.go:117] "RemoveContainer" containerID="f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.142746 4837 scope.go:117] "RemoveContainer" containerID="c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.149244 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.168596 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d98fd957f-h677b"] Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.173302 4837 scope.go:117] "RemoveContainer" containerID="f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1" Oct 01 08:47:03 crc kubenswrapper[4837]: E1001 08:47:03.174128 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1\": container with ID starting with f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1 not found: ID does not exist" containerID="f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.174181 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1"} err="failed to get container status \"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1\": rpc error: code = NotFound desc = could not find container \"f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1\": container with ID starting with f8b41caf7c2b7b8b33844f93c9d152cac5616a78f25b659dac25b84eaa8d02a1 not found: ID does not exist" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.174219 4837 scope.go:117] "RemoveContainer" containerID="c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5" Oct 01 08:47:03 crc kubenswrapper[4837]: E1001 08:47:03.174627 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5\": container with ID starting with c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5 not found: ID does not exist" containerID="c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.174665 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5"} err="failed to get container status \"c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5\": rpc error: code = NotFound desc = could not find container \"c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5\": container with ID starting with c3d25da1a10754f57cf9f8c5979862a673dc85a554a183d43b7eefe4bb8f24b5 not found: ID does not exist" Oct 01 08:47:03 crc kubenswrapper[4837]: I1001 08:47:03.832493 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" path="/var/lib/kubelet/pods/eda09a1e-e8ee-4af5-99b3-54601be231d1/volumes" Oct 01 08:47:23 crc kubenswrapper[4837]: I1001 08:47:23.080367 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:47:23 crc kubenswrapper[4837]: I1001 08:47:23.081150 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:47:32 crc kubenswrapper[4837]: I1001 08:47:32.378272 4837 generic.go:334] "Generic (PLEG): container finished" podID="b86b0277-4772-4869-93dc-e261cd56b06f" containerID="1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe" exitCode=0 Oct 01 08:47:32 crc kubenswrapper[4837]: I1001 08:47:32.378387 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerDied","Data":"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe"} Oct 01 08:47:32 crc kubenswrapper[4837]: I1001 08:47:32.384360 4837 generic.go:334] "Generic (PLEG): container finished" podID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerID="de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5" exitCode=0 Oct 01 08:47:32 crc kubenswrapper[4837]: I1001 08:47:32.384432 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerDied","Data":"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5"} Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.395571 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerStarted","Data":"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f"} Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.396906 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.399643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerStarted","Data":"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb"} Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.399957 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.438990 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.012770626 podStartE2EDuration="1m22.438954105s" podCreationTimestamp="2025-10-01 08:46:11 +0000 UTC" firstStartedPulling="2025-10-01 08:46:13.781007086 +0000 UTC m=+6030.622614541" lastFinishedPulling="2025-10-01 08:46:57.207190565 +0000 UTC m=+6074.048798020" observedRunningTime="2025-10-01 08:47:33.430578949 +0000 UTC m=+6110.272186414" watchObservedRunningTime="2025-10-01 08:47:33.438954105 +0000 UTC m=+6110.280561560" Oct 01 08:47:33 crc kubenswrapper[4837]: I1001 08:47:33.476973 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.663792176 podStartE2EDuration="1m22.476955689s" podCreationTimestamp="2025-10-01 08:46:11 +0000 UTC" firstStartedPulling="2025-10-01 08:46:13.369027498 +0000 UTC m=+6030.210634943" lastFinishedPulling="2025-10-01 08:46:57.182191001 +0000 UTC m=+6074.023798456" observedRunningTime="2025-10-01 08:47:33.474823166 +0000 UTC m=+6110.316430621" watchObservedRunningTime="2025-10-01 08:47:33.476955689 +0000 UTC m=+6110.318563144" Oct 01 08:47:42 crc kubenswrapper[4837]: I1001 08:47:42.877065 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.251:5671: connect: connection refused" Oct 01 08:47:43 crc kubenswrapper[4837]: I1001 08:47:43.218115 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:47:52 crc kubenswrapper[4837]: I1001 08:47:52.875954 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 08:47:53 crc kubenswrapper[4837]: I1001 08:47:53.080152 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:47:53 crc kubenswrapper[4837]: I1001 08:47:53.080234 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.952332 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953760 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="dnsmasq-dns" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953779 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="dnsmasq-dns" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953803 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953810 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953822 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="extract-content" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953829 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="extract-content" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953843 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1083379-0b1d-4003-81e4-6cc8d9839478" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953850 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1083379-0b1d-4003-81e4-6cc8d9839478" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953880 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="extract-utilities" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953890 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="extract-utilities" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953906 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c33f13b-00b7-48ae-9151-b77dfe90ad21" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953912 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c33f13b-00b7-48ae-9151-b77dfe90ad21" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: E1001 08:47:55.953923 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="registry-server" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.953930 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="registry-server" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.954119 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c33f13b-00b7-48ae-9151-b77dfe90ad21" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.954144 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda09a1e-e8ee-4af5-99b3-54601be231d1" containerName="dnsmasq-dns" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.954154 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59f1c24-1509-415c-a312-c4d5e9b02f6c" containerName="registry-server" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.954173 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1083379-0b1d-4003-81e4-6cc8d9839478" containerName="init" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.955979 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:55 crc kubenswrapper[4837]: I1001 08:47:55.972272 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.118350 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62x6\" (UniqueName: \"kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.118490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.118556 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.220065 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.220183 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.220251 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62x6\" (UniqueName: \"kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.222400 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.222824 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.255209 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62x6\" (UniqueName: \"kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6\") pod \"dnsmasq-dns-84966d4785-xjgcs\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.282554 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.690588 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:47:56 crc kubenswrapper[4837]: I1001 08:47:56.876025 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:47:57 crc kubenswrapper[4837]: I1001 08:47:57.613033 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:47:57 crc kubenswrapper[4837]: I1001 08:47:57.656949 4837 generic.go:334] "Generic (PLEG): container finished" podID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerID="5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6" exitCode=0 Oct 01 08:47:57 crc kubenswrapper[4837]: I1001 08:47:57.657007 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" event={"ID":"9275d07b-d7b2-4d5e-bee1-296a9115b49a","Type":"ContainerDied","Data":"5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6"} Oct 01 08:47:57 crc kubenswrapper[4837]: I1001 08:47:57.657050 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" event={"ID":"9275d07b-d7b2-4d5e-bee1-296a9115b49a","Type":"ContainerStarted","Data":"0de9561a6bec652e89f65f83b71088fd894d8b4c465d012595040d576282705c"} Oct 01 08:47:58 crc kubenswrapper[4837]: I1001 08:47:58.670391 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" event={"ID":"9275d07b-d7b2-4d5e-bee1-296a9115b49a","Type":"ContainerStarted","Data":"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00"} Oct 01 08:47:58 crc kubenswrapper[4837]: I1001 08:47:58.672124 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:48:01 crc kubenswrapper[4837]: I1001 08:48:01.932448 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" containerID="cri-o://62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb" gracePeriod=604795 Oct 01 08:48:02 crc kubenswrapper[4837]: I1001 08:48:02.746561 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="rabbitmq" containerID="cri-o://9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f" gracePeriod=604795 Oct 01 08:48:02 crc kubenswrapper[4837]: I1001 08:48:02.875610 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.251:5671: connect: connection refused" Oct 01 08:48:03 crc kubenswrapper[4837]: I1001 08:48:03.215805 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.252:5671: connect: connection refused" Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.284984 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.306578 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" podStartSLOduration=11.30654941 podStartE2EDuration="11.30654941s" podCreationTimestamp="2025-10-01 08:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:47:58.700217466 +0000 UTC m=+6135.541824961" watchObservedRunningTime="2025-10-01 08:48:06.30654941 +0000 UTC m=+6143.148156865" Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.340057 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.340361 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-689c49c455-7kv79" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="dnsmasq-dns" containerID="cri-o://c7140bb7a671a4a474642202d44fd4b2fa5c4a4818b629fe055a94e36fc9a360" gracePeriod=10 Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.755477 4837 generic.go:334] "Generic (PLEG): container finished" podID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerID="c7140bb7a671a4a474642202d44fd4b2fa5c4a4818b629fe055a94e36fc9a360" exitCode=0 Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.755531 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689c49c455-7kv79" event={"ID":"0be6f3e3-91de-4d71-abe7-d1b2beed5476","Type":"ContainerDied","Data":"c7140bb7a671a4a474642202d44fd4b2fa5c4a4818b629fe055a94e36fc9a360"} Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.755884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689c49c455-7kv79" event={"ID":"0be6f3e3-91de-4d71-abe7-d1b2beed5476","Type":"ContainerDied","Data":"e0a67a0abe38859944070c94fadd32a984a466b1894a4dfcb57bc29e5ee4bb6d"} Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.755905 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0a67a0abe38859944070c94fadd32a984a466b1894a4dfcb57bc29e5ee4bb6d" Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.793193 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.942191 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srzj4\" (UniqueName: \"kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4\") pod \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.942314 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config\") pod \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.942346 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc\") pod \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\" (UID: \"0be6f3e3-91de-4d71-abe7-d1b2beed5476\") " Oct 01 08:48:06 crc kubenswrapper[4837]: I1001 08:48:06.950716 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4" (OuterVolumeSpecName: "kube-api-access-srzj4") pod "0be6f3e3-91de-4d71-abe7-d1b2beed5476" (UID: "0be6f3e3-91de-4d71-abe7-d1b2beed5476"). InnerVolumeSpecName "kube-api-access-srzj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.016891 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config" (OuterVolumeSpecName: "config") pod "0be6f3e3-91de-4d71-abe7-d1b2beed5476" (UID: "0be6f3e3-91de-4d71-abe7-d1b2beed5476"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.019272 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0be6f3e3-91de-4d71-abe7-d1b2beed5476" (UID: "0be6f3e3-91de-4d71-abe7-d1b2beed5476"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.043874 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srzj4\" (UniqueName: \"kubernetes.io/projected/0be6f3e3-91de-4d71-abe7-d1b2beed5476-kube-api-access-srzj4\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.043911 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.043921 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0be6f3e3-91de-4d71-abe7-d1b2beed5476-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.763527 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689c49c455-7kv79" Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.792708 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.803001 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-689c49c455-7kv79"] Oct 01 08:48:07 crc kubenswrapper[4837]: I1001 08:48:07.825997 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" path="/var/lib/kubelet/pods/0be6f3e3-91de-4d71-abe7-d1b2beed5476/volumes" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.626050 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.775514 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.775662 4837 generic.go:334] "Generic (PLEG): container finished" podID="b86b0277-4772-4869-93dc-e261cd56b06f" containerID="62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb" exitCode=0 Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.775714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerDied","Data":"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb"} Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.777375 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b86b0277-4772-4869-93dc-e261cd56b06f","Type":"ContainerDied","Data":"c8c05fed940c5ef91b66e412eae28328a6a3f2365f5fb1aeda6df34842413e36"} Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.775794 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.777503 4837 scope.go:117] "RemoveContainer" containerID="62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.782334 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.783720 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.784228 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.784562 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.784734 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.785102 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.785728 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.786598 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.786767 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhkfx\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.786902 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.787046 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.788193 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls\") pod \"b86b0277-4772-4869-93dc-e261cd56b06f\" (UID: \"b86b0277-4772-4869-93dc-e261cd56b06f\") " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.789962 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info" (OuterVolumeSpecName: "pod-info") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.790797 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.790985 4837 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.791068 4837 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b86b0277-4772-4869-93dc-e261cd56b06f-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.793415 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx" (OuterVolumeSpecName: "kube-api-access-lhkfx") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "kube-api-access-lhkfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.793571 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.794073 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.807928 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.811052 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c" (OuterVolumeSpecName: "persistence") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.828098 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data" (OuterVolumeSpecName: "config-data") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.855668 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf" (OuterVolumeSpecName: "server-conf") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895317 4837 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b86b0277-4772-4869-93dc-e261cd56b06f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895425 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") on node \"crc\" " Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895455 4837 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895473 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhkfx\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-kube-api-access-lhkfx\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895489 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b86b0277-4772-4869-93dc-e261cd56b06f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895499 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.895510 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.900821 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b86b0277-4772-4869-93dc-e261cd56b06f" (UID: "b86b0277-4772-4869-93dc-e261cd56b06f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.920955 4837 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.921312 4837 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c") on node "crc" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.997152 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b86b0277-4772-4869-93dc-e261cd56b06f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:08 crc kubenswrapper[4837]: I1001 08:48:08.997199 4837 reconciler_common.go:293] "Volume detached for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.026007 4837 scope.go:117] "RemoveContainer" containerID="1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.065786 4837 scope.go:117] "RemoveContainer" containerID="62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.074084 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb\": container with ID starting with 62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb not found: ID does not exist" containerID="62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.074170 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb"} err="failed to get container status \"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb\": rpc error: code = NotFound desc = could not find container \"62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb\": container with ID starting with 62b0478a66a6d57acb662b9157d8edfe727ae62eff35c80b8153e753499293bb not found: ID does not exist" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.074212 4837 scope.go:117] "RemoveContainer" containerID="1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.075470 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe\": container with ID starting with 1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe not found: ID does not exist" containerID="1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.075516 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe"} err="failed to get container status \"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe\": rpc error: code = NotFound desc = could not find container \"1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe\": container with ID starting with 1f4028b2053afac5cc967d5d1b11b50efb0b982a08a81c27756349d1ddc744fe not found: ID does not exist" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.123383 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.129305 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.159464 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.159856 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="dnsmasq-dns" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.159879 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="dnsmasq-dns" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.159903 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="setup-container" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.159913 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="setup-container" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.159930 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.159939 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.159963 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="init" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.159971 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="init" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.160174 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be6f3e3-91de-4d71-abe7-d1b2beed5476" containerName="dnsmasq-dns" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.160193 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.161207 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.192087 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.192203 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.192436 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.198124 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.198338 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.198447 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.198559 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ts9d5" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.199860 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.301926 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cda9191-47f1-4a0e-9c7a-32775705c5a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302056 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302102 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302155 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302176 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302204 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302222 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302387 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfhr\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-kube-api-access-9tfhr\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302457 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cda9191-47f1-4a0e-9c7a-32775705c5a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.302492 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.343774 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.408447 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.408523 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cda9191-47f1-4a0e-9c7a-32775705c5a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.408582 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.409648 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cda9191-47f1-4a0e-9c7a-32775705c5a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.409817 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.409889 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410014 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410062 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410151 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfhr\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-kube-api-access-9tfhr\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.410744 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.411090 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.411908 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.412345 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.416202 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.416276 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afbb63deeecbf25cf80875790908ff7c42a9c13794df34babbfb4f9f2a1e87b2/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.417057 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cda9191-47f1-4a0e-9c7a-32775705c5a5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.417639 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.417717 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.418323 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cda9191-47f1-4a0e-9c7a-32775705c5a5-config-data\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.421160 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cda9191-47f1-4a0e-9c7a-32775705c5a5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.432708 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfhr\" (UniqueName: \"kubernetes.io/projected/3cda9191-47f1-4a0e-9c7a-32775705c5a5-kube-api-access-9tfhr\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.460113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c7b3b76-89ee-4327-a90d-25d21ed0787c\") pod \"rabbitmq-server-0\" (UID: \"3cda9191-47f1-4a0e-9c7a-32775705c5a5\") " pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.483831 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511382 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511449 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511569 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511629 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511649 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gxvv\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511685 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511841 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511886 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511910 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511944 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf\") pod \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\" (UID: \"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7\") " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.511971 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.512264 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.512316 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.512907 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.517621 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv" (OuterVolumeSpecName: "kube-api-access-9gxvv") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "kube-api-access-9gxvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.517952 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.520921 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info" (OuterVolumeSpecName: "pod-info") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.520919 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.540311 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372" (OuterVolumeSpecName: "persistence") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.581554 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data" (OuterVolumeSpecName: "config-data") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.603715 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf" (OuterVolumeSpecName: "server-conf") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616564 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") on node \"crc\" " Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616594 4837 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616606 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gxvv\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-kube-api-access-9gxvv\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616616 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616624 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616634 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616646 4837 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616670 4837 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.616678 4837 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.640794 4837 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.641515 4837 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372") on node "crc" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.720995 4837 reconciler_common.go:293] "Volume detached for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.762946 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" (UID: "c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.823987 4837 generic.go:334] "Generic (PLEG): container finished" podID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerID="9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f" exitCode=0 Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.824097 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.834881 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b86b0277-4772-4869-93dc-e261cd56b06f" path="/var/lib/kubelet/pods/b86b0277-4772-4869-93dc-e261cd56b06f/volumes" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.835189 4837 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.837302 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerDied","Data":"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f"} Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.837363 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7","Type":"ContainerDied","Data":"925cefbc3ed7cb0c13912c708203da969027054d9502461f422c128745da3336"} Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.837405 4837 scope.go:117] "RemoveContainer" containerID="9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.888733 4837 scope.go:117] "RemoveContainer" containerID="de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.900609 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.904740 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.938438 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.938907 4837 scope.go:117] "RemoveContainer" containerID="9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.939630 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="setup-container" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.939656 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="setup-container" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.939688 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.939713 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.943126 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" containerName="rabbitmq" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.943803 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f\": container with ID starting with 9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f not found: ID does not exist" containerID="9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.943969 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f"} err="failed to get container status \"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f\": rpc error: code = NotFound desc = could not find container \"9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f\": container with ID starting with 9e41f8f20547f986c8953a472fd22e6a49afb9181327496b051588ff6f5f826f not found: ID does not exist" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.944114 4837 scope.go:117] "RemoveContainer" containerID="de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5" Oct 01 08:48:09 crc kubenswrapper[4837]: E1001 08:48:09.945034 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5\": container with ID starting with de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5 not found: ID does not exist" containerID="de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.945088 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.945091 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5"} err="failed to get container status \"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5\": rpc error: code = NotFound desc = could not find container \"de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5\": container with ID starting with de6f896f469bf9d4078a371096d54d466131ac500be1f377204e2d993aa062f5 not found: ID does not exist" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.949685 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.949777 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.950413 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.950478 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.950413 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.951197 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-htgzk" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.956645 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 08:48:09 crc kubenswrapper[4837]: I1001 08:48:09.975492 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.096326 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146743 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cf357d6-04ea-4f07-a430-32eaba72e2da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146868 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cf357d6-04ea-4f07-a430-32eaba72e2da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146947 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.146988 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjnck\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-kube-api-access-tjnck\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.147003 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.147041 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.147181 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.248491 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.248541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.248569 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cf357d6-04ea-4f07-a430-32eaba72e2da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.248586 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.248647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249329 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249372 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjnck\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-kube-api-access-tjnck\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249393 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249423 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249458 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249491 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cf357d6-04ea-4f07-a430-32eaba72e2da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249907 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.249925 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.251164 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.251664 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.253008 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cf357d6-04ea-4f07-a430-32eaba72e2da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.253872 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cf357d6-04ea-4f07-a430-32eaba72e2da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.253956 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cf357d6-04ea-4f07-a430-32eaba72e2da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.255810 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.255876 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ad0c11f22fa60721e63eac5a1d2193d37d85e125324ebb604a4e9e68815f42a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.257851 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.260462 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.269855 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjnck\" (UniqueName: \"kubernetes.io/projected/0cf357d6-04ea-4f07-a430-32eaba72e2da-kube-api-access-tjnck\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.294972 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8b6f037b-038d-4fda-9eb6-8ddc282e1372\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cf357d6-04ea-4f07-a430-32eaba72e2da\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.585444 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.847797 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cda9191-47f1-4a0e-9c7a-32775705c5a5","Type":"ContainerStarted","Data":"c4a2c2c8453be85b6607b62fba414c734d008639ff585756fb064aec7a01f59f"} Oct 01 08:48:10 crc kubenswrapper[4837]: I1001 08:48:10.996206 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 08:48:11 crc kubenswrapper[4837]: I1001 08:48:11.829526 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7" path="/var/lib/kubelet/pods/c216a3dd-cce4-4c8a-89c4-7f65c1ba63f7/volumes" Oct 01 08:48:11 crc kubenswrapper[4837]: I1001 08:48:11.856003 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cf357d6-04ea-4f07-a430-32eaba72e2da","Type":"ContainerStarted","Data":"60f85f63caaa79a59183bf25a694c77e8ac4881f72dbc90c02779acd749521d0"} Oct 01 08:48:11 crc kubenswrapper[4837]: I1001 08:48:11.859067 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cda9191-47f1-4a0e-9c7a-32775705c5a5","Type":"ContainerStarted","Data":"a7c277fbec56eee58eaa0b7bc76272524f2f89d7937e1685ab2004b761fb1df6"} Oct 01 08:48:12 crc kubenswrapper[4837]: I1001 08:48:12.870214 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cf357d6-04ea-4f07-a430-32eaba72e2da","Type":"ContainerStarted","Data":"e8710f5f289342c3409e0372a45478ca5d62181a0706bc722f1a2578ca987f8e"} Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.079461 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.080093 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.080156 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.080932 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.081033 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada" gracePeriod=600 Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.987079 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada" exitCode=0 Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.987183 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada"} Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.987470 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921"} Oct 01 08:48:23 crc kubenswrapper[4837]: I1001 08:48:23.987498 4837 scope.go:117] "RemoveContainer" containerID="0028e0ac2fa43f5ac915cd6da0e990f2a13f922cd5adaf0fa2abef99ddcf4088" Oct 01 08:48:44 crc kubenswrapper[4837]: I1001 08:48:44.190226 4837 generic.go:334] "Generic (PLEG): container finished" podID="3cda9191-47f1-4a0e-9c7a-32775705c5a5" containerID="a7c277fbec56eee58eaa0b7bc76272524f2f89d7937e1685ab2004b761fb1df6" exitCode=0 Oct 01 08:48:44 crc kubenswrapper[4837]: I1001 08:48:44.190372 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cda9191-47f1-4a0e-9c7a-32775705c5a5","Type":"ContainerDied","Data":"a7c277fbec56eee58eaa0b7bc76272524f2f89d7937e1685ab2004b761fb1df6"} Oct 01 08:48:45 crc kubenswrapper[4837]: I1001 08:48:45.205228 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3cda9191-47f1-4a0e-9c7a-32775705c5a5","Type":"ContainerStarted","Data":"6c933f57da41da9acbcccd8c8d174a5502527a77a1c7236ff498d1317f92f493"} Oct 01 08:48:45 crc kubenswrapper[4837]: I1001 08:48:45.205921 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 08:48:45 crc kubenswrapper[4837]: I1001 08:48:45.242247 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.242220796 podStartE2EDuration="36.242220796s" podCreationTimestamp="2025-10-01 08:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:48:45.233009289 +0000 UTC m=+6182.074616744" watchObservedRunningTime="2025-10-01 08:48:45.242220796 +0000 UTC m=+6182.083828291" Oct 01 08:48:46 crc kubenswrapper[4837]: I1001 08:48:46.216892 4837 generic.go:334] "Generic (PLEG): container finished" podID="0cf357d6-04ea-4f07-a430-32eaba72e2da" containerID="e8710f5f289342c3409e0372a45478ca5d62181a0706bc722f1a2578ca987f8e" exitCode=0 Oct 01 08:48:46 crc kubenswrapper[4837]: I1001 08:48:46.217521 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cf357d6-04ea-4f07-a430-32eaba72e2da","Type":"ContainerDied","Data":"e8710f5f289342c3409e0372a45478ca5d62181a0706bc722f1a2578ca987f8e"} Oct 01 08:48:47 crc kubenswrapper[4837]: I1001 08:48:47.228679 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cf357d6-04ea-4f07-a430-32eaba72e2da","Type":"ContainerStarted","Data":"5ada93681335db1fe472f8bcbae06b6349d61f98a0dc841cb1565b5f9a3fc9c7"} Oct 01 08:48:47 crc kubenswrapper[4837]: I1001 08:48:47.229269 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:48:47 crc kubenswrapper[4837]: I1001 08:48:47.256249 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.256222061 podStartE2EDuration="38.256222061s" podCreationTimestamp="2025-10-01 08:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:48:47.252604312 +0000 UTC m=+6184.094211807" watchObservedRunningTime="2025-10-01 08:48:47.256222061 +0000 UTC m=+6184.097829536" Oct 01 08:48:59 crc kubenswrapper[4837]: I1001 08:48:59.486979 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 08:49:00 crc kubenswrapper[4837]: I1001 08:49:00.589870 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.420669 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.424855 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.431831 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cz4sn" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.440353 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.559247 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5km84\" (UniqueName: \"kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84\") pod \"mariadb-client-1-default\" (UID: \"5fba28dc-842a-47f2-b3ef-b0e73c39b83a\") " pod="openstack/mariadb-client-1-default" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.660741 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5km84\" (UniqueName: \"kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84\") pod \"mariadb-client-1-default\" (UID: \"5fba28dc-842a-47f2-b3ef-b0e73c39b83a\") " pod="openstack/mariadb-client-1-default" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.683083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5km84\" (UniqueName: \"kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84\") pod \"mariadb-client-1-default\" (UID: \"5fba28dc-842a-47f2-b3ef-b0e73c39b83a\") " pod="openstack/mariadb-client-1-default" Oct 01 08:49:02 crc kubenswrapper[4837]: I1001 08:49:02.762539 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 01 08:49:03 crc kubenswrapper[4837]: I1001 08:49:03.347103 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 01 08:49:03 crc kubenswrapper[4837]: I1001 08:49:03.413268 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"5fba28dc-842a-47f2-b3ef-b0e73c39b83a","Type":"ContainerStarted","Data":"34b51300105d821d5299206fee9ccb61d981c36d49ad8472d890d033ceeac139"} Oct 01 08:49:04 crc kubenswrapper[4837]: I1001 08:49:04.429000 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fba28dc-842a-47f2-b3ef-b0e73c39b83a" containerID="e26efa6503c3fca6b612eeb16087628aa7800d085f50c97b1d4a59821e17867d" exitCode=0 Oct 01 08:49:04 crc kubenswrapper[4837]: I1001 08:49:04.429128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"5fba28dc-842a-47f2-b3ef-b0e73c39b83a","Type":"ContainerDied","Data":"e26efa6503c3fca6b612eeb16087628aa7800d085f50c97b1d4a59821e17867d"} Oct 01 08:49:05 crc kubenswrapper[4837]: I1001 08:49:05.919077 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 01 08:49:05 crc kubenswrapper[4837]: I1001 08:49:05.949558 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_5fba28dc-842a-47f2-b3ef-b0e73c39b83a/mariadb-client-1-default/0.log" Oct 01 08:49:05 crc kubenswrapper[4837]: I1001 08:49:05.980633 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 01 08:49:05 crc kubenswrapper[4837]: I1001 08:49:05.989849 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.019601 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5km84\" (UniqueName: \"kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84\") pod \"5fba28dc-842a-47f2-b3ef-b0e73c39b83a\" (UID: \"5fba28dc-842a-47f2-b3ef-b0e73c39b83a\") " Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.025646 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84" (OuterVolumeSpecName: "kube-api-access-5km84") pod "5fba28dc-842a-47f2-b3ef-b0e73c39b83a" (UID: "5fba28dc-842a-47f2-b3ef-b0e73c39b83a"). InnerVolumeSpecName "kube-api-access-5km84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.123052 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5km84\" (UniqueName: \"kubernetes.io/projected/5fba28dc-842a-47f2-b3ef-b0e73c39b83a-kube-api-access-5km84\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.427947 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 01 08:49:06 crc kubenswrapper[4837]: E1001 08:49:06.428667 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fba28dc-842a-47f2-b3ef-b0e73c39b83a" containerName="mariadb-client-1-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.428711 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fba28dc-842a-47f2-b3ef-b0e73c39b83a" containerName="mariadb-client-1-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.428918 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fba28dc-842a-47f2-b3ef-b0e73c39b83a" containerName="mariadb-client-1-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.429624 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.435478 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.462745 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34b51300105d821d5299206fee9ccb61d981c36d49ad8472d890d033ceeac139" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.462845 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.530761 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbnt\" (UniqueName: \"kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt\") pod \"mariadb-client-2-default\" (UID: \"291a2eea-f1ad-441f-8510-4143b7e4acfc\") " pod="openstack/mariadb-client-2-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.632748 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbnt\" (UniqueName: \"kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt\") pod \"mariadb-client-2-default\" (UID: \"291a2eea-f1ad-441f-8510-4143b7e4acfc\") " pod="openstack/mariadb-client-2-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.665533 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbnt\" (UniqueName: \"kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt\") pod \"mariadb-client-2-default\" (UID: \"291a2eea-f1ad-441f-8510-4143b7e4acfc\") " pod="openstack/mariadb-client-2-default" Oct 01 08:49:06 crc kubenswrapper[4837]: I1001 08:49:06.769347 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 01 08:49:07 crc kubenswrapper[4837]: I1001 08:49:07.410215 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 01 08:49:07 crc kubenswrapper[4837]: W1001 08:49:07.416249 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod291a2eea_f1ad_441f_8510_4143b7e4acfc.slice/crio-42032d1c1a49c77db5568a893b8bd21009ae3672806ff2a39f461bf2a7023941 WatchSource:0}: Error finding container 42032d1c1a49c77db5568a893b8bd21009ae3672806ff2a39f461bf2a7023941: Status 404 returned error can't find the container with id 42032d1c1a49c77db5568a893b8bd21009ae3672806ff2a39f461bf2a7023941 Oct 01 08:49:07 crc kubenswrapper[4837]: I1001 08:49:07.473131 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"291a2eea-f1ad-441f-8510-4143b7e4acfc","Type":"ContainerStarted","Data":"42032d1c1a49c77db5568a893b8bd21009ae3672806ff2a39f461bf2a7023941"} Oct 01 08:49:07 crc kubenswrapper[4837]: I1001 08:49:07.827287 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fba28dc-842a-47f2-b3ef-b0e73c39b83a" path="/var/lib/kubelet/pods/5fba28dc-842a-47f2-b3ef-b0e73c39b83a/volumes" Oct 01 08:49:08 crc kubenswrapper[4837]: I1001 08:49:08.493328 4837 generic.go:334] "Generic (PLEG): container finished" podID="291a2eea-f1ad-441f-8510-4143b7e4acfc" containerID="abc94b8067b10a8c31b17162c4a758a9edf1757d265a15aeb2eaa2d86e703387" exitCode=0 Oct 01 08:49:08 crc kubenswrapper[4837]: I1001 08:49:08.493464 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"291a2eea-f1ad-441f-8510-4143b7e4acfc","Type":"ContainerDied","Data":"abc94b8067b10a8c31b17162c4a758a9edf1757d265a15aeb2eaa2d86e703387"} Oct 01 08:49:09 crc kubenswrapper[4837]: I1001 08:49:09.955568 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.021334 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_291a2eea-f1ad-441f-8510-4143b7e4acfc/mariadb-client-2-default/0.log" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.054115 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.060887 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.101337 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbnt\" (UniqueName: \"kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt\") pod \"291a2eea-f1ad-441f-8510-4143b7e4acfc\" (UID: \"291a2eea-f1ad-441f-8510-4143b7e4acfc\") " Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.110912 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt" (OuterVolumeSpecName: "kube-api-access-kwbnt") pod "291a2eea-f1ad-441f-8510-4143b7e4acfc" (UID: "291a2eea-f1ad-441f-8510-4143b7e4acfc"). InnerVolumeSpecName "kube-api-access-kwbnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.203347 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbnt\" (UniqueName: \"kubernetes.io/projected/291a2eea-f1ad-441f-8510-4143b7e4acfc-kube-api-access-kwbnt\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.236159 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-3-default"] Oct 01 08:49:10 crc kubenswrapper[4837]: E1001 08:49:10.236614 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="291a2eea-f1ad-441f-8510-4143b7e4acfc" containerName="mariadb-client-2-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.236636 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="291a2eea-f1ad-441f-8510-4143b7e4acfc" containerName="mariadb-client-2-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.236839 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="291a2eea-f1ad-441f-8510-4143b7e4acfc" containerName="mariadb-client-2-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.237653 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.248007 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.408594 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmmrh\" (UniqueName: \"kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh\") pod \"mariadb-client-3-default\" (UID: \"ab1f03c3-62ee-4128-b0de-2138177d9369\") " pod="openstack/mariadb-client-3-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.510944 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmmrh\" (UniqueName: \"kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh\") pod \"mariadb-client-3-default\" (UID: \"ab1f03c3-62ee-4128-b0de-2138177d9369\") " pod="openstack/mariadb-client-3-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.514275 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42032d1c1a49c77db5568a893b8bd21009ae3672806ff2a39f461bf2a7023941" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.514386 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.532387 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmmrh\" (UniqueName: \"kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh\") pod \"mariadb-client-3-default\" (UID: \"ab1f03c3-62ee-4128-b0de-2138177d9369\") " pod="openstack/mariadb-client-3-default" Oct 01 08:49:10 crc kubenswrapper[4837]: I1001 08:49:10.560839 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 01 08:49:11 crc kubenswrapper[4837]: I1001 08:49:11.002575 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 01 08:49:11 crc kubenswrapper[4837]: W1001 08:49:11.014186 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab1f03c3_62ee_4128_b0de_2138177d9369.slice/crio-e68de92be1a26f8f2e463b71da7811ac3d8cdf8d3d012fc8369425890d981bb2 WatchSource:0}: Error finding container e68de92be1a26f8f2e463b71da7811ac3d8cdf8d3d012fc8369425890d981bb2: Status 404 returned error can't find the container with id e68de92be1a26f8f2e463b71da7811ac3d8cdf8d3d012fc8369425890d981bb2 Oct 01 08:49:11 crc kubenswrapper[4837]: I1001 08:49:11.528350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"ab1f03c3-62ee-4128-b0de-2138177d9369","Type":"ContainerStarted","Data":"f7f2246bfd375936eb2c2b4bd287dfb8c62e101d4b835d7b07d0eb77b18d8d8a"} Oct 01 08:49:11 crc kubenswrapper[4837]: I1001 08:49:11.528433 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"ab1f03c3-62ee-4128-b0de-2138177d9369","Type":"ContainerStarted","Data":"e68de92be1a26f8f2e463b71da7811ac3d8cdf8d3d012fc8369425890d981bb2"} Oct 01 08:49:11 crc kubenswrapper[4837]: I1001 08:49:11.559730 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-3-default" podStartSLOduration=1.559673184 podStartE2EDuration="1.559673184s" podCreationTimestamp="2025-10-01 08:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:49:11.550769546 +0000 UTC m=+6208.392377041" watchObservedRunningTime="2025-10-01 08:49:11.559673184 +0000 UTC m=+6208.401280669" Oct 01 08:49:11 crc kubenswrapper[4837]: I1001 08:49:11.833630 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="291a2eea-f1ad-441f-8510-4143b7e4acfc" path="/var/lib/kubelet/pods/291a2eea-f1ad-441f-8510-4143b7e4acfc/volumes" Oct 01 08:49:13 crc kubenswrapper[4837]: I1001 08:49:13.550790 4837 generic.go:334] "Generic (PLEG): container finished" podID="ab1f03c3-62ee-4128-b0de-2138177d9369" containerID="f7f2246bfd375936eb2c2b4bd287dfb8c62e101d4b835d7b07d0eb77b18d8d8a" exitCode=0 Oct 01 08:49:13 crc kubenswrapper[4837]: I1001 08:49:13.550896 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"ab1f03c3-62ee-4128-b0de-2138177d9369","Type":"ContainerDied","Data":"f7f2246bfd375936eb2c2b4bd287dfb8c62e101d4b835d7b07d0eb77b18d8d8a"} Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.023648 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.064196 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.070286 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.210842 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmmrh\" (UniqueName: \"kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh\") pod \"ab1f03c3-62ee-4128-b0de-2138177d9369\" (UID: \"ab1f03c3-62ee-4128-b0de-2138177d9369\") " Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.220056 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh" (OuterVolumeSpecName: "kube-api-access-fmmrh") pod "ab1f03c3-62ee-4128-b0de-2138177d9369" (UID: "ab1f03c3-62ee-4128-b0de-2138177d9369"). InnerVolumeSpecName "kube-api-access-fmmrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.313737 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmmrh\" (UniqueName: \"kubernetes.io/projected/ab1f03c3-62ee-4128-b0de-2138177d9369-kube-api-access-fmmrh\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.542493 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 01 08:49:15 crc kubenswrapper[4837]: E1001 08:49:15.543356 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1f03c3-62ee-4128-b0de-2138177d9369" containerName="mariadb-client-3-default" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.543382 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1f03c3-62ee-4128-b0de-2138177d9369" containerName="mariadb-client-3-default" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.543604 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab1f03c3-62ee-4128-b0de-2138177d9369" containerName="mariadb-client-3-default" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.544254 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.558819 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.571159 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68de92be1a26f8f2e463b71da7811ac3d8cdf8d3d012fc8369425890d981bb2" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.571257 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.720200 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbv74\" (UniqueName: \"kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74\") pod \"mariadb-client-1\" (UID: \"c2ca62d0-a359-438f-b3e1-326d153851f0\") " pod="openstack/mariadb-client-1" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.822109 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbv74\" (UniqueName: \"kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74\") pod \"mariadb-client-1\" (UID: \"c2ca62d0-a359-438f-b3e1-326d153851f0\") " pod="openstack/mariadb-client-1" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.834985 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab1f03c3-62ee-4128-b0de-2138177d9369" path="/var/lib/kubelet/pods/ab1f03c3-62ee-4128-b0de-2138177d9369/volumes" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.843571 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbv74\" (UniqueName: \"kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74\") pod \"mariadb-client-1\" (UID: \"c2ca62d0-a359-438f-b3e1-326d153851f0\") " pod="openstack/mariadb-client-1" Oct 01 08:49:15 crc kubenswrapper[4837]: I1001 08:49:15.874189 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 01 08:49:16 crc kubenswrapper[4837]: I1001 08:49:16.438361 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 01 08:49:16 crc kubenswrapper[4837]: I1001 08:49:16.580481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"c2ca62d0-a359-438f-b3e1-326d153851f0","Type":"ContainerStarted","Data":"c25151aefca171361c7959cd4a7aeaba02f32ecc6c29479394368dd9a9ec8b9d"} Oct 01 08:49:17 crc kubenswrapper[4837]: I1001 08:49:17.594994 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2ca62d0-a359-438f-b3e1-326d153851f0" containerID="28a44c948369728e9fa4e1a95e8fc5715453e79653300fb842adcced46acc7b0" exitCode=0 Oct 01 08:49:17 crc kubenswrapper[4837]: I1001 08:49:17.595041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"c2ca62d0-a359-438f-b3e1-326d153851f0","Type":"ContainerDied","Data":"28a44c948369728e9fa4e1a95e8fc5715453e79653300fb842adcced46acc7b0"} Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.061456 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.083404 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_c2ca62d0-a359-438f-b3e1-326d153851f0/mariadb-client-1/0.log" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.113746 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.121343 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.183564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbv74\" (UniqueName: \"kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74\") pod \"c2ca62d0-a359-438f-b3e1-326d153851f0\" (UID: \"c2ca62d0-a359-438f-b3e1-326d153851f0\") " Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.192042 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74" (OuterVolumeSpecName: "kube-api-access-wbv74") pod "c2ca62d0-a359-438f-b3e1-326d153851f0" (UID: "c2ca62d0-a359-438f-b3e1-326d153851f0"). InnerVolumeSpecName "kube-api-access-wbv74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.286053 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbv74\" (UniqueName: \"kubernetes.io/projected/c2ca62d0-a359-438f-b3e1-326d153851f0-kube-api-access-wbv74\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.582657 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 01 08:49:19 crc kubenswrapper[4837]: E1001 08:49:19.583214 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ca62d0-a359-438f-b3e1-326d153851f0" containerName="mariadb-client-1" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.583249 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ca62d0-a359-438f-b3e1-326d153851f0" containerName="mariadb-client-1" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.583565 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ca62d0-a359-438f-b3e1-326d153851f0" containerName="mariadb-client-1" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.584401 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.595348 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.619285 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c25151aefca171361c7959cd4a7aeaba02f32ecc6c29479394368dd9a9ec8b9d" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.619389 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.694187 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrllb\" (UniqueName: \"kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb\") pod \"mariadb-client-4-default\" (UID: \"095281ca-f99f-4aba-aba6-e6f9a7a5b112\") " pod="openstack/mariadb-client-4-default" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.795715 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrllb\" (UniqueName: \"kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb\") pod \"mariadb-client-4-default\" (UID: \"095281ca-f99f-4aba-aba6-e6f9a7a5b112\") " pod="openstack/mariadb-client-4-default" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.815116 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrllb\" (UniqueName: \"kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb\") pod \"mariadb-client-4-default\" (UID: \"095281ca-f99f-4aba-aba6-e6f9a7a5b112\") " pod="openstack/mariadb-client-4-default" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.826491 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ca62d0-a359-438f-b3e1-326d153851f0" path="/var/lib/kubelet/pods/c2ca62d0-a359-438f-b3e1-326d153851f0/volumes" Oct 01 08:49:19 crc kubenswrapper[4837]: I1001 08:49:19.915463 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 01 08:49:20 crc kubenswrapper[4837]: I1001 08:49:20.247848 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 01 08:49:20 crc kubenswrapper[4837]: I1001 08:49:20.631448 4837 generic.go:334] "Generic (PLEG): container finished" podID="095281ca-f99f-4aba-aba6-e6f9a7a5b112" containerID="f62599eb7b29368b067c586f5143bcffe3072109bc10d7d76e1452c533489606" exitCode=0 Oct 01 08:49:20 crc kubenswrapper[4837]: I1001 08:49:20.631510 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"095281ca-f99f-4aba-aba6-e6f9a7a5b112","Type":"ContainerDied","Data":"f62599eb7b29368b067c586f5143bcffe3072109bc10d7d76e1452c533489606"} Oct 01 08:49:20 crc kubenswrapper[4837]: I1001 08:49:20.631549 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"095281ca-f99f-4aba-aba6-e6f9a7a5b112","Type":"ContainerStarted","Data":"77ff787426be5a763ab830eae4e893519aa51da598f94a6667902acbad7f2e78"} Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.126076 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.129936 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.148606 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.329868 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.330007 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.330230 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7tjw\" (UniqueName: \"kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.432640 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.432733 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.432866 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7tjw\" (UniqueName: \"kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.433392 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.433549 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.467973 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7tjw\" (UniqueName: \"kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw\") pod \"certified-operators-7p2b7\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:21 crc kubenswrapper[4837]: I1001 08:49:21.481269 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.034223 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.093024 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.166751 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_095281ca-f99f-4aba-aba6-e6f9a7a5b112/mariadb-client-4-default/0.log" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.199167 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.204053 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.269433 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrllb\" (UniqueName: \"kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb\") pod \"095281ca-f99f-4aba-aba6-e6f9a7a5b112\" (UID: \"095281ca-f99f-4aba-aba6-e6f9a7a5b112\") " Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.276893 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb" (OuterVolumeSpecName: "kube-api-access-zrllb") pod "095281ca-f99f-4aba-aba6-e6f9a7a5b112" (UID: "095281ca-f99f-4aba-aba6-e6f9a7a5b112"). InnerVolumeSpecName "kube-api-access-zrllb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.371935 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrllb\" (UniqueName: \"kubernetes.io/projected/095281ca-f99f-4aba-aba6-e6f9a7a5b112-kube-api-access-zrllb\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.658271 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77ff787426be5a763ab830eae4e893519aa51da598f94a6667902acbad7f2e78" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.658293 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.661311 4837 generic.go:334] "Generic (PLEG): container finished" podID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerID="c0ab491e7d095b13abbdc3b77287fe2c95874191068f41edd7bd49aff22c1c69" exitCode=0 Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.661362 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerDied","Data":"c0ab491e7d095b13abbdc3b77287fe2c95874191068f41edd7bd49aff22c1c69"} Oct 01 08:49:22 crc kubenswrapper[4837]: I1001 08:49:22.661396 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerStarted","Data":"c9a5660a6798d369bc3e86fbc807d1518920509e5c0e1989cc4c46e20be2947f"} Oct 01 08:49:23 crc kubenswrapper[4837]: I1001 08:49:23.673459 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerStarted","Data":"30e43dd3fb9549ff5da33b2f252e949a5bc5414ace447b3118dc77f8cbaba256"} Oct 01 08:49:23 crc kubenswrapper[4837]: I1001 08:49:23.834380 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="095281ca-f99f-4aba-aba6-e6f9a7a5b112" path="/var/lib/kubelet/pods/095281ca-f99f-4aba-aba6-e6f9a7a5b112/volumes" Oct 01 08:49:24 crc kubenswrapper[4837]: I1001 08:49:24.690834 4837 generic.go:334] "Generic (PLEG): container finished" podID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerID="30e43dd3fb9549ff5da33b2f252e949a5bc5414ace447b3118dc77f8cbaba256" exitCode=0 Oct 01 08:49:24 crc kubenswrapper[4837]: I1001 08:49:24.690899 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerDied","Data":"30e43dd3fb9549ff5da33b2f252e949a5bc5414ace447b3118dc77f8cbaba256"} Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.712897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerStarted","Data":"3659ed8fcbe2ed8a8d8afe220c8c051f7c90bd36b989bd545335bb5c17457350"} Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.744752 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7p2b7" podStartSLOduration=2.24694489 podStartE2EDuration="4.74473171s" podCreationTimestamp="2025-10-01 08:49:21 +0000 UTC" firstStartedPulling="2025-10-01 08:49:22.663775634 +0000 UTC m=+6219.505383089" lastFinishedPulling="2025-10-01 08:49:25.161562454 +0000 UTC m=+6222.003169909" observedRunningTime="2025-10-01 08:49:25.744009123 +0000 UTC m=+6222.585616588" watchObservedRunningTime="2025-10-01 08:49:25.74473171 +0000 UTC m=+6222.586339165" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.760867 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 01 08:49:25 crc kubenswrapper[4837]: E1001 08:49:25.761322 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095281ca-f99f-4aba-aba6-e6f9a7a5b112" containerName="mariadb-client-4-default" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.761345 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="095281ca-f99f-4aba-aba6-e6f9a7a5b112" containerName="mariadb-client-4-default" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.761564 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="095281ca-f99f-4aba-aba6-e6f9a7a5b112" containerName="mariadb-client-4-default" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.762340 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.770147 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cz4sn" Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.778801 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 01 08:49:25 crc kubenswrapper[4837]: I1001 08:49:25.941938 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd5x9\" (UniqueName: \"kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9\") pod \"mariadb-client-5-default\" (UID: \"75f0156b-373b-4b9c-8f77-4900487fd641\") " pod="openstack/mariadb-client-5-default" Oct 01 08:49:26 crc kubenswrapper[4837]: I1001 08:49:26.043829 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd5x9\" (UniqueName: \"kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9\") pod \"mariadb-client-5-default\" (UID: \"75f0156b-373b-4b9c-8f77-4900487fd641\") " pod="openstack/mariadb-client-5-default" Oct 01 08:49:26 crc kubenswrapper[4837]: I1001 08:49:26.071270 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd5x9\" (UniqueName: \"kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9\") pod \"mariadb-client-5-default\" (UID: \"75f0156b-373b-4b9c-8f77-4900487fd641\") " pod="openstack/mariadb-client-5-default" Oct 01 08:49:26 crc kubenswrapper[4837]: I1001 08:49:26.084729 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 01 08:49:26 crc kubenswrapper[4837]: I1001 08:49:26.635952 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 01 08:49:26 crc kubenswrapper[4837]: I1001 08:49:26.722393 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"75f0156b-373b-4b9c-8f77-4900487fd641","Type":"ContainerStarted","Data":"c55239b75d73f9a842f03e7cef9c3fc463cd40dbfc6d01aa7f5e9f8112e813b0"} Oct 01 08:49:27 crc kubenswrapper[4837]: I1001 08:49:27.735259 4837 generic.go:334] "Generic (PLEG): container finished" podID="75f0156b-373b-4b9c-8f77-4900487fd641" containerID="120985a71b11d26ddedf5a3e596de0b8c5e907937ccb4965ccbf0e1705f6609b" exitCode=0 Oct 01 08:49:27 crc kubenswrapper[4837]: I1001 08:49:27.735307 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"75f0156b-373b-4b9c-8f77-4900487fd641","Type":"ContainerDied","Data":"120985a71b11d26ddedf5a3e596de0b8c5e907937ccb4965ccbf0e1705f6609b"} Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.222251 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.247144 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_75f0156b-373b-4b9c-8f77-4900487fd641/mariadb-client-5-default/0.log" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.281258 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.289291 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.402137 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd5x9\" (UniqueName: \"kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9\") pod \"75f0156b-373b-4b9c-8f77-4900487fd641\" (UID: \"75f0156b-373b-4b9c-8f77-4900487fd641\") " Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.416101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9" (OuterVolumeSpecName: "kube-api-access-rd5x9") pod "75f0156b-373b-4b9c-8f77-4900487fd641" (UID: "75f0156b-373b-4b9c-8f77-4900487fd641"). InnerVolumeSpecName "kube-api-access-rd5x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.419944 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 01 08:49:29 crc kubenswrapper[4837]: E1001 08:49:29.420506 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f0156b-373b-4b9c-8f77-4900487fd641" containerName="mariadb-client-5-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.420529 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f0156b-373b-4b9c-8f77-4900487fd641" containerName="mariadb-client-5-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.421522 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f0156b-373b-4b9c-8f77-4900487fd641" containerName="mariadb-client-5-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.422362 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.429937 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.504393 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd5x9\" (UniqueName: \"kubernetes.io/projected/75f0156b-373b-4b9c-8f77-4900487fd641-kube-api-access-rd5x9\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.605647 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbq9g\" (UniqueName: \"kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g\") pod \"mariadb-client-6-default\" (UID: \"82181a1d-7669-4540-9e4a-e78c101df192\") " pod="openstack/mariadb-client-6-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.707751 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbq9g\" (UniqueName: \"kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g\") pod \"mariadb-client-6-default\" (UID: \"82181a1d-7669-4540-9e4a-e78c101df192\") " pod="openstack/mariadb-client-6-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.739322 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbq9g\" (UniqueName: \"kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g\") pod \"mariadb-client-6-default\" (UID: \"82181a1d-7669-4540-9e4a-e78c101df192\") " pod="openstack/mariadb-client-6-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.761496 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.765740 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55239b75d73f9a842f03e7cef9c3fc463cd40dbfc6d01aa7f5e9f8112e813b0" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.765851 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 01 08:49:29 crc kubenswrapper[4837]: I1001 08:49:29.828991 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f0156b-373b-4b9c-8f77-4900487fd641" path="/var/lib/kubelet/pods/75f0156b-373b-4b9c-8f77-4900487fd641/volumes" Oct 01 08:49:30 crc kubenswrapper[4837]: I1001 08:49:30.365246 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 01 08:49:30 crc kubenswrapper[4837]: I1001 08:49:30.779065 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"82181a1d-7669-4540-9e4a-e78c101df192","Type":"ContainerStarted","Data":"0f0940c4cc010aa437bca66772184b4b82d0b3f4ce39a89128d0bf2c062a7299"} Oct 01 08:49:30 crc kubenswrapper[4837]: I1001 08:49:30.779115 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"82181a1d-7669-4540-9e4a-e78c101df192","Type":"ContainerStarted","Data":"8ff14457861191e8f53bc46b5e634e6302c0c66b1fe125c184acf3f6d3fd43dd"} Oct 01 08:49:30 crc kubenswrapper[4837]: I1001 08:49:30.805747 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.8057263479999999 podStartE2EDuration="1.805726348s" podCreationTimestamp="2025-10-01 08:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:49:30.797722391 +0000 UTC m=+6227.639329836" watchObservedRunningTime="2025-10-01 08:49:30.805726348 +0000 UTC m=+6227.647333803" Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.482831 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.483225 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.532675 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.791618 4837 generic.go:334] "Generic (PLEG): container finished" podID="82181a1d-7669-4540-9e4a-e78c101df192" containerID="0f0940c4cc010aa437bca66772184b4b82d0b3f4ce39a89128d0bf2c062a7299" exitCode=0 Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.791686 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"82181a1d-7669-4540-9e4a-e78c101df192","Type":"ContainerDied","Data":"0f0940c4cc010aa437bca66772184b4b82d0b3f4ce39a89128d0bf2c062a7299"} Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.861793 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:31 crc kubenswrapper[4837]: I1001 08:49:31.933010 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.292665 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.336736 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.341023 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.483272 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbq9g\" (UniqueName: \"kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g\") pod \"82181a1d-7669-4540-9e4a-e78c101df192\" (UID: \"82181a1d-7669-4540-9e4a-e78c101df192\") " Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.492328 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 01 08:49:33 crc kubenswrapper[4837]: E1001 08:49:33.493153 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82181a1d-7669-4540-9e4a-e78c101df192" containerName="mariadb-client-6-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.493172 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="82181a1d-7669-4540-9e4a-e78c101df192" containerName="mariadb-client-6-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.494534 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="82181a1d-7669-4540-9e4a-e78c101df192" containerName="mariadb-client-6-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.495546 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.495784 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g" (OuterVolumeSpecName: "kube-api-access-mbq9g") pod "82181a1d-7669-4540-9e4a-e78c101df192" (UID: "82181a1d-7669-4540-9e4a-e78c101df192"). InnerVolumeSpecName "kube-api-access-mbq9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.509834 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.585316 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbq9g\" (UniqueName: \"kubernetes.io/projected/82181a1d-7669-4540-9e4a-e78c101df192-kube-api-access-mbq9g\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.687390 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr599\" (UniqueName: \"kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599\") pod \"mariadb-client-7-default\" (UID: \"3e25a19e-3011-40c4-9b1b-57dab48237a8\") " pod="openstack/mariadb-client-7-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.790426 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr599\" (UniqueName: \"kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599\") pod \"mariadb-client-7-default\" (UID: \"3e25a19e-3011-40c4-9b1b-57dab48237a8\") " pod="openstack/mariadb-client-7-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.816821 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr599\" (UniqueName: \"kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599\") pod \"mariadb-client-7-default\" (UID: \"3e25a19e-3011-40c4-9b1b-57dab48237a8\") " pod="openstack/mariadb-client-7-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.826060 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.826238 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7p2b7" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="registry-server" containerID="cri-o://3659ed8fcbe2ed8a8d8afe220c8c051f7c90bd36b989bd545335bb5c17457350" gracePeriod=2 Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.841184 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82181a1d-7669-4540-9e4a-e78c101df192" path="/var/lib/kubelet/pods/82181a1d-7669-4540-9e4a-e78c101df192/volumes" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.842601 4837 scope.go:117] "RemoveContainer" containerID="0f0940c4cc010aa437bca66772184b4b82d0b3f4ce39a89128d0bf2c062a7299" Oct 01 08:49:33 crc kubenswrapper[4837]: I1001 08:49:33.867904 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.454660 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 01 08:49:34 crc kubenswrapper[4837]: W1001 08:49:34.463375 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e25a19e_3011_40c4_9b1b_57dab48237a8.slice/crio-3d63e01d6215b5fbe6d387af5d8a81424a8c860922d9a64b9f9fc02a8fa73609 WatchSource:0}: Error finding container 3d63e01d6215b5fbe6d387af5d8a81424a8c860922d9a64b9f9fc02a8fa73609: Status 404 returned error can't find the container with id 3d63e01d6215b5fbe6d387af5d8a81424a8c860922d9a64b9f9fc02a8fa73609 Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.832486 4837 generic.go:334] "Generic (PLEG): container finished" podID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerID="3659ed8fcbe2ed8a8d8afe220c8c051f7c90bd36b989bd545335bb5c17457350" exitCode=0 Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.832594 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerDied","Data":"3659ed8fcbe2ed8a8d8afe220c8c051f7c90bd36b989bd545335bb5c17457350"} Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.836584 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e25a19e-3011-40c4-9b1b-57dab48237a8" containerID="831293fe9fd86d206b4e31dd5d26f9afb31f57575c205c6cf44e4067133dece5" exitCode=0 Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.836625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3e25a19e-3011-40c4-9b1b-57dab48237a8","Type":"ContainerDied","Data":"831293fe9fd86d206b4e31dd5d26f9afb31f57575c205c6cf44e4067133dece5"} Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.836643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3e25a19e-3011-40c4-9b1b-57dab48237a8","Type":"ContainerStarted","Data":"3d63e01d6215b5fbe6d387af5d8a81424a8c860922d9a64b9f9fc02a8fa73609"} Oct 01 08:49:34 crc kubenswrapper[4837]: I1001 08:49:34.952161 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.115460 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities\") pod \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.115844 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content\") pod \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.116028 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7tjw\" (UniqueName: \"kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw\") pod \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\" (UID: \"fa1cd8df-4c24-407b-a95e-448bdc02d0d2\") " Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.120009 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities" (OuterVolumeSpecName: "utilities") pod "fa1cd8df-4c24-407b-a95e-448bdc02d0d2" (UID: "fa1cd8df-4c24-407b-a95e-448bdc02d0d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.122950 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw" (OuterVolumeSpecName: "kube-api-access-v7tjw") pod "fa1cd8df-4c24-407b-a95e-448bdc02d0d2" (UID: "fa1cd8df-4c24-407b-a95e-448bdc02d0d2"). InnerVolumeSpecName "kube-api-access-v7tjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.199958 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa1cd8df-4c24-407b-a95e-448bdc02d0d2" (UID: "fa1cd8df-4c24-407b-a95e-448bdc02d0d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.218821 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.218901 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.218930 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7tjw\" (UniqueName: \"kubernetes.io/projected/fa1cd8df-4c24-407b-a95e-448bdc02d0d2-kube-api-access-v7tjw\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.852308 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7p2b7" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.852465 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7p2b7" event={"ID":"fa1cd8df-4c24-407b-a95e-448bdc02d0d2","Type":"ContainerDied","Data":"c9a5660a6798d369bc3e86fbc807d1518920509e5c0e1989cc4c46e20be2947f"} Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.852589 4837 scope.go:117] "RemoveContainer" containerID="3659ed8fcbe2ed8a8d8afe220c8c051f7c90bd36b989bd545335bb5c17457350" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.900634 4837 scope.go:117] "RemoveContainer" containerID="30e43dd3fb9549ff5da33b2f252e949a5bc5414ace447b3118dc77f8cbaba256" Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.902327 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.915766 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7p2b7"] Oct 01 08:49:35 crc kubenswrapper[4837]: I1001 08:49:35.947703 4837 scope.go:117] "RemoveContainer" containerID="c0ab491e7d095b13abbdc3b77287fe2c95874191068f41edd7bd49aff22c1c69" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.354554 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.374194 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_3e25a19e-3011-40c4-9b1b-57dab48237a8/mariadb-client-7-default/0.log" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.408925 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.414005 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.448552 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr599\" (UniqueName: \"kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599\") pod \"3e25a19e-3011-40c4-9b1b-57dab48237a8\" (UID: \"3e25a19e-3011-40c4-9b1b-57dab48237a8\") " Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.453857 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599" (OuterVolumeSpecName: "kube-api-access-jr599") pod "3e25a19e-3011-40c4-9b1b-57dab48237a8" (UID: "3e25a19e-3011-40c4-9b1b-57dab48237a8"). InnerVolumeSpecName "kube-api-access-jr599". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528032 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 01 08:49:36 crc kubenswrapper[4837]: E1001 08:49:36.528361 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="extract-content" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528375 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="extract-content" Oct 01 08:49:36 crc kubenswrapper[4837]: E1001 08:49:36.528389 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="extract-utilities" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528395 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="extract-utilities" Oct 01 08:49:36 crc kubenswrapper[4837]: E1001 08:49:36.528406 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e25a19e-3011-40c4-9b1b-57dab48237a8" containerName="mariadb-client-7-default" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528413 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e25a19e-3011-40c4-9b1b-57dab48237a8" containerName="mariadb-client-7-default" Oct 01 08:49:36 crc kubenswrapper[4837]: E1001 08:49:36.528431 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="registry-server" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528437 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="registry-server" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528599 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e25a19e-3011-40c4-9b1b-57dab48237a8" containerName="mariadb-client-7-default" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.528619 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" containerName="registry-server" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.529770 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.537461 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.550483 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr599\" (UniqueName: \"kubernetes.io/projected/3e25a19e-3011-40c4-9b1b-57dab48237a8-kube-api-access-jr599\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.651733 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v47w4\" (UniqueName: \"kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4\") pod \"mariadb-client-2\" (UID: \"263c517f-cdcc-4a49-91af-0451ea9604eb\") " pod="openstack/mariadb-client-2" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.753133 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v47w4\" (UniqueName: \"kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4\") pod \"mariadb-client-2\" (UID: \"263c517f-cdcc-4a49-91af-0451ea9604eb\") " pod="openstack/mariadb-client-2" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.775332 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v47w4\" (UniqueName: \"kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4\") pod \"mariadb-client-2\" (UID: \"263c517f-cdcc-4a49-91af-0451ea9604eb\") " pod="openstack/mariadb-client-2" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.856644 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.868152 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d63e01d6215b5fbe6d387af5d8a81424a8c860922d9a64b9f9fc02a8fa73609" Oct 01 08:49:36 crc kubenswrapper[4837]: I1001 08:49:36.868269 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.428017 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 01 08:49:37 crc kubenswrapper[4837]: W1001 08:49:37.434683 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod263c517f_cdcc_4a49_91af_0451ea9604eb.slice/crio-18548d44c52a554987fefc1d7b1d7f59c7cb81e28ae6f869d77ae964fb5224af WatchSource:0}: Error finding container 18548d44c52a554987fefc1d7b1d7f59c7cb81e28ae6f869d77ae964fb5224af: Status 404 returned error can't find the container with id 18548d44c52a554987fefc1d7b1d7f59c7cb81e28ae6f869d77ae964fb5224af Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.829663 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e25a19e-3011-40c4-9b1b-57dab48237a8" path="/var/lib/kubelet/pods/3e25a19e-3011-40c4-9b1b-57dab48237a8/volumes" Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.830632 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa1cd8df-4c24-407b-a95e-448bdc02d0d2" path="/var/lib/kubelet/pods/fa1cd8df-4c24-407b-a95e-448bdc02d0d2/volumes" Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.879824 4837 generic.go:334] "Generic (PLEG): container finished" podID="263c517f-cdcc-4a49-91af-0451ea9604eb" containerID="53eea99fe6947e1cb5bb478e13fd9564430e7f3622139325006bb181ad499a0c" exitCode=0 Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.879890 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"263c517f-cdcc-4a49-91af-0451ea9604eb","Type":"ContainerDied","Data":"53eea99fe6947e1cb5bb478e13fd9564430e7f3622139325006bb181ad499a0c"} Oct 01 08:49:37 crc kubenswrapper[4837]: I1001 08:49:37.879934 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"263c517f-cdcc-4a49-91af-0451ea9604eb","Type":"ContainerStarted","Data":"18548d44c52a554987fefc1d7b1d7f59c7cb81e28ae6f869d77ae964fb5224af"} Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.365920 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.392265 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_263c517f-cdcc-4a49-91af-0451ea9604eb/mariadb-client-2/0.log" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.420261 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.429992 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.502308 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47w4\" (UniqueName: \"kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4\") pod \"263c517f-cdcc-4a49-91af-0451ea9604eb\" (UID: \"263c517f-cdcc-4a49-91af-0451ea9604eb\") " Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.509194 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4" (OuterVolumeSpecName: "kube-api-access-v47w4") pod "263c517f-cdcc-4a49-91af-0451ea9604eb" (UID: "263c517f-cdcc-4a49-91af-0451ea9604eb"). InnerVolumeSpecName "kube-api-access-v47w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.604911 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47w4\" (UniqueName: \"kubernetes.io/projected/263c517f-cdcc-4a49-91af-0451ea9604eb-kube-api-access-v47w4\") on node \"crc\" DevicePath \"\"" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.825493 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263c517f-cdcc-4a49-91af-0451ea9604eb" path="/var/lib/kubelet/pods/263c517f-cdcc-4a49-91af-0451ea9604eb/volumes" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.904061 4837 scope.go:117] "RemoveContainer" containerID="53eea99fe6947e1cb5bb478e13fd9564430e7f3622139325006bb181ad499a0c" Oct 01 08:49:39 crc kubenswrapper[4837]: I1001 08:49:39.904209 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 01 08:50:05 crc kubenswrapper[4837]: I1001 08:50:05.666925 4837 scope.go:117] "RemoveContainer" containerID="9eaf6a53d9ce56cab43fb1e811b1916ba25dd9eac1509d9536edd5f6f601e89c" Oct 01 08:50:23 crc kubenswrapper[4837]: I1001 08:50:23.079264 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:50:23 crc kubenswrapper[4837]: I1001 08:50:23.081054 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:50:53 crc kubenswrapper[4837]: I1001 08:50:53.079860 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:50:53 crc kubenswrapper[4837]: I1001 08:50:53.081992 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.079508 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.080078 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.080122 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.080831 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.080897 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" gracePeriod=600 Oct 01 08:51:23 crc kubenswrapper[4837]: E1001 08:51:23.219756 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.920760 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" exitCode=0 Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.921093 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921"} Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.921216 4837 scope.go:117] "RemoveContainer" containerID="16d3f916206ef8215ec4b9d251a92b07b2f5fe01f2928a9b74f84b8894cc5ada" Oct 01 08:51:23 crc kubenswrapper[4837]: I1001 08:51:23.921767 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:51:23 crc kubenswrapper[4837]: E1001 08:51:23.922051 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:51:37 crc kubenswrapper[4837]: I1001 08:51:37.816260 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:51:37 crc kubenswrapper[4837]: E1001 08:51:37.817341 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:51:51 crc kubenswrapper[4837]: I1001 08:51:51.818544 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:51:51 crc kubenswrapper[4837]: E1001 08:51:51.819648 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:52:03 crc kubenswrapper[4837]: I1001 08:52:03.821883 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:52:03 crc kubenswrapper[4837]: E1001 08:52:03.824027 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:52:17 crc kubenswrapper[4837]: I1001 08:52:17.816998 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:52:17 crc kubenswrapper[4837]: E1001 08:52:17.818006 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:52:31 crc kubenswrapper[4837]: I1001 08:52:31.816306 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:52:31 crc kubenswrapper[4837]: E1001 08:52:31.817438 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:52:44 crc kubenswrapper[4837]: I1001 08:52:44.817279 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:52:44 crc kubenswrapper[4837]: E1001 08:52:44.821157 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:52:57 crc kubenswrapper[4837]: I1001 08:52:57.816837 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:52:57 crc kubenswrapper[4837]: E1001 08:52:57.817887 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:53:05 crc kubenswrapper[4837]: I1001 08:53:05.830849 4837 scope.go:117] "RemoveContainer" containerID="f1350c91350dba84024084a37f54eff5fa68ddfeabb695465b639d4fc3036fbd" Oct 01 08:53:05 crc kubenswrapper[4837]: I1001 08:53:05.864689 4837 scope.go:117] "RemoveContainer" containerID="c7140bb7a671a4a474642202d44fd4b2fa5c4a4818b629fe055a94e36fc9a360" Oct 01 08:53:11 crc kubenswrapper[4837]: I1001 08:53:11.816631 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:53:11 crc kubenswrapper[4837]: E1001 08:53:11.817843 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:53:23 crc kubenswrapper[4837]: I1001 08:53:23.821953 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:53:23 crc kubenswrapper[4837]: E1001 08:53:23.823404 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:53:37 crc kubenswrapper[4837]: I1001 08:53:37.816020 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:53:37 crc kubenswrapper[4837]: E1001 08:53:37.816749 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:53:52 crc kubenswrapper[4837]: I1001 08:53:52.816394 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:53:52 crc kubenswrapper[4837]: E1001 08:53:52.817127 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:54:05 crc kubenswrapper[4837]: I1001 08:54:05.816621 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:54:05 crc kubenswrapper[4837]: E1001 08:54:05.817713 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:54:19 crc kubenswrapper[4837]: I1001 08:54:19.816121 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:54:19 crc kubenswrapper[4837]: E1001 08:54:19.816957 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:54:34 crc kubenswrapper[4837]: I1001 08:54:34.816905 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:54:34 crc kubenswrapper[4837]: E1001 08:54:34.817934 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:54:46 crc kubenswrapper[4837]: I1001 08:54:46.816433 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:54:46 crc kubenswrapper[4837]: E1001 08:54:46.818107 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.152808 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:54:51 crc kubenswrapper[4837]: E1001 08:54:51.153836 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263c517f-cdcc-4a49-91af-0451ea9604eb" containerName="mariadb-client-2" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.153855 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="263c517f-cdcc-4a49-91af-0451ea9604eb" containerName="mariadb-client-2" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.154024 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="263c517f-cdcc-4a49-91af-0451ea9604eb" containerName="mariadb-client-2" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.155502 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.171761 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.233626 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.233678 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.233744 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pkkk\" (UniqueName: \"kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.335438 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.335561 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pkkk\" (UniqueName: \"kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.335640 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.336118 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.336390 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.357925 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pkkk\" (UniqueName: \"kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk\") pod \"redhat-operators-5gv7p\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.474171 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:54:51 crc kubenswrapper[4837]: I1001 08:54:51.923441 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:54:52 crc kubenswrapper[4837]: I1001 08:54:52.921655 4837 generic.go:334] "Generic (PLEG): container finished" podID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerID="27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03" exitCode=0 Oct 01 08:54:52 crc kubenswrapper[4837]: I1001 08:54:52.921949 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerDied","Data":"27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03"} Oct 01 08:54:52 crc kubenswrapper[4837]: I1001 08:54:52.921978 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerStarted","Data":"21fb071a9c8959be7c59f23719294fbcdb5ff6355f7552a0f59f0744270eaf1d"} Oct 01 08:54:52 crc kubenswrapper[4837]: I1001 08:54:52.924404 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.555542 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.557314 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.571275 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.579014 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qvpf\" (UniqueName: \"kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.579095 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.579173 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.680890 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qvpf\" (UniqueName: \"kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.680942 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.680984 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.681424 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.681459 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.700706 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qvpf\" (UniqueName: \"kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf\") pod \"community-operators-vl7vv\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:53 crc kubenswrapper[4837]: I1001 08:54:53.880652 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.465242 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.947124 4837 generic.go:334] "Generic (PLEG): container finished" podID="56410990-174e-4fd5-94e7-e7ef93eae654" containerID="f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd" exitCode=0 Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.947213 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerDied","Data":"f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd"} Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.947250 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerStarted","Data":"f27fd7d6aad7ee7620d90984b7b8fc0a4c52d5342d07be1a29b2849188a770dd"} Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.950846 4837 generic.go:334] "Generic (PLEG): container finished" podID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerID="44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e" exitCode=0 Oct 01 08:54:54 crc kubenswrapper[4837]: I1001 08:54:54.950888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerDied","Data":"44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e"} Oct 01 08:54:55 crc kubenswrapper[4837]: I1001 08:54:55.980262 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerStarted","Data":"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1"} Oct 01 08:54:56 crc kubenswrapper[4837]: I1001 08:54:56.991501 4837 generic.go:334] "Generic (PLEG): container finished" podID="56410990-174e-4fd5-94e7-e7ef93eae654" containerID="4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea" exitCode=0 Oct 01 08:54:56 crc kubenswrapper[4837]: I1001 08:54:56.991574 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerDied","Data":"4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea"} Oct 01 08:54:57 crc kubenswrapper[4837]: I1001 08:54:57.016598 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5gv7p" podStartSLOduration=3.4607493639999998 podStartE2EDuration="6.016566936s" podCreationTimestamp="2025-10-01 08:54:51 +0000 UTC" firstStartedPulling="2025-10-01 08:54:52.924035164 +0000 UTC m=+6549.765642639" lastFinishedPulling="2025-10-01 08:54:55.479852746 +0000 UTC m=+6552.321460211" observedRunningTime="2025-10-01 08:54:56.013913686 +0000 UTC m=+6552.855521151" watchObservedRunningTime="2025-10-01 08:54:57.016566936 +0000 UTC m=+6553.858174411" Oct 01 08:54:58 crc kubenswrapper[4837]: I1001 08:54:58.005317 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerStarted","Data":"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57"} Oct 01 08:54:58 crc kubenswrapper[4837]: I1001 08:54:58.031057 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vl7vv" podStartSLOduration=2.477850977 podStartE2EDuration="5.031037215s" podCreationTimestamp="2025-10-01 08:54:53 +0000 UTC" firstStartedPulling="2025-10-01 08:54:54.949426317 +0000 UTC m=+6551.791033802" lastFinishedPulling="2025-10-01 08:54:57.502612585 +0000 UTC m=+6554.344220040" observedRunningTime="2025-10-01 08:54:58.02547862 +0000 UTC m=+6554.867086105" watchObservedRunningTime="2025-10-01 08:54:58.031037215 +0000 UTC m=+6554.872644670" Oct 01 08:54:58 crc kubenswrapper[4837]: I1001 08:54:58.816609 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:54:58 crc kubenswrapper[4837]: E1001 08:54:58.817319 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:55:01 crc kubenswrapper[4837]: I1001 08:55:01.474888 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:01 crc kubenswrapper[4837]: I1001 08:55:01.475268 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:01 crc kubenswrapper[4837]: I1001 08:55:01.558088 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:02 crc kubenswrapper[4837]: I1001 08:55:02.109974 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:02 crc kubenswrapper[4837]: I1001 08:55:02.174269 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:55:03 crc kubenswrapper[4837]: I1001 08:55:03.881198 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:03 crc kubenswrapper[4837]: I1001 08:55:03.881622 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:03 crc kubenswrapper[4837]: I1001 08:55:03.956829 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.064990 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5gv7p" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="registry-server" containerID="cri-o://5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1" gracePeriod=2 Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.121060 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.518124 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.542894 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.699180 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content\") pod \"a65db51c-f0f3-48e6-908f-8b34321296b1\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.699300 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities\") pod \"a65db51c-f0f3-48e6-908f-8b34321296b1\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.699402 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pkkk\" (UniqueName: \"kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk\") pod \"a65db51c-f0f3-48e6-908f-8b34321296b1\" (UID: \"a65db51c-f0f3-48e6-908f-8b34321296b1\") " Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.701357 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities" (OuterVolumeSpecName: "utilities") pod "a65db51c-f0f3-48e6-908f-8b34321296b1" (UID: "a65db51c-f0f3-48e6-908f-8b34321296b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.708514 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk" (OuterVolumeSpecName: "kube-api-access-4pkkk") pod "a65db51c-f0f3-48e6-908f-8b34321296b1" (UID: "a65db51c-f0f3-48e6-908f-8b34321296b1"). InnerVolumeSpecName "kube-api-access-4pkkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.803120 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pkkk\" (UniqueName: \"kubernetes.io/projected/a65db51c-f0f3-48e6-908f-8b34321296b1-kube-api-access-4pkkk\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:04 crc kubenswrapper[4837]: I1001 08:55:04.803167 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.077409 4837 generic.go:334] "Generic (PLEG): container finished" podID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerID="5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1" exitCode=0 Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.077518 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5gv7p" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.077519 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerDied","Data":"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1"} Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.077599 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5gv7p" event={"ID":"a65db51c-f0f3-48e6-908f-8b34321296b1","Type":"ContainerDied","Data":"21fb071a9c8959be7c59f23719294fbcdb5ff6355f7552a0f59f0744270eaf1d"} Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.077630 4837 scope.go:117] "RemoveContainer" containerID="5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.110055 4837 scope.go:117] "RemoveContainer" containerID="44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.155572 4837 scope.go:117] "RemoveContainer" containerID="27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.189807 4837 scope.go:117] "RemoveContainer" containerID="5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1" Oct 01 08:55:05 crc kubenswrapper[4837]: E1001 08:55:05.190158 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1\": container with ID starting with 5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1 not found: ID does not exist" containerID="5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.190191 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1"} err="failed to get container status \"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1\": rpc error: code = NotFound desc = could not find container \"5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1\": container with ID starting with 5c25809f4abe4925f491757a2860bcf501226c075b20efce0c9e43038ca377d1 not found: ID does not exist" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.190219 4837 scope.go:117] "RemoveContainer" containerID="44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e" Oct 01 08:55:05 crc kubenswrapper[4837]: E1001 08:55:05.190442 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e\": container with ID starting with 44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e not found: ID does not exist" containerID="44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.190475 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e"} err="failed to get container status \"44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e\": rpc error: code = NotFound desc = could not find container \"44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e\": container with ID starting with 44ca34eec2de634fc38d8d72976fbfbbcd3ea8ec56d794bf3eda1f4ff469d51e not found: ID does not exist" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.190495 4837 scope.go:117] "RemoveContainer" containerID="27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03" Oct 01 08:55:05 crc kubenswrapper[4837]: E1001 08:55:05.190800 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03\": container with ID starting with 27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03 not found: ID does not exist" containerID="27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.190854 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03"} err="failed to get container status \"27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03\": rpc error: code = NotFound desc = could not find container \"27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03\": container with ID starting with 27ca252b79fea87a596e40c8fc6a4db0ef98cfddb787be2154e7a6504c1c9f03 not found: ID does not exist" Oct 01 08:55:05 crc kubenswrapper[4837]: I1001 08:55:05.959085 4837 scope.go:117] "RemoveContainer" containerID="e26efa6503c3fca6b612eeb16087628aa7800d085f50c97b1d4a59821e17867d" Oct 01 08:55:06 crc kubenswrapper[4837]: I1001 08:55:06.014223 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a65db51c-f0f3-48e6-908f-8b34321296b1" (UID: "a65db51c-f0f3-48e6-908f-8b34321296b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:55:06 crc kubenswrapper[4837]: I1001 08:55:06.024115 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a65db51c-f0f3-48e6-908f-8b34321296b1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:06 crc kubenswrapper[4837]: I1001 08:55:06.089962 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vl7vv" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="registry-server" containerID="cri-o://5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57" gracePeriod=2 Oct 01 08:55:06 crc kubenswrapper[4837]: I1001 08:55:06.309792 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:55:06 crc kubenswrapper[4837]: I1001 08:55:06.315479 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5gv7p"] Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.010063 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.103380 4837 generic.go:334] "Generic (PLEG): container finished" podID="56410990-174e-4fd5-94e7-e7ef93eae654" containerID="5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57" exitCode=0 Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.103445 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerDied","Data":"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57"} Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.103493 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vl7vv" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.103555 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vl7vv" event={"ID":"56410990-174e-4fd5-94e7-e7ef93eae654","Type":"ContainerDied","Data":"f27fd7d6aad7ee7620d90984b7b8fc0a4c52d5342d07be1a29b2849188a770dd"} Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.103600 4837 scope.go:117] "RemoveContainer" containerID="5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.131920 4837 scope.go:117] "RemoveContainer" containerID="4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.143556 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities\") pod \"56410990-174e-4fd5-94e7-e7ef93eae654\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.143646 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content\") pod \"56410990-174e-4fd5-94e7-e7ef93eae654\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.143810 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qvpf\" (UniqueName: \"kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf\") pod \"56410990-174e-4fd5-94e7-e7ef93eae654\" (UID: \"56410990-174e-4fd5-94e7-e7ef93eae654\") " Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.145291 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities" (OuterVolumeSpecName: "utilities") pod "56410990-174e-4fd5-94e7-e7ef93eae654" (UID: "56410990-174e-4fd5-94e7-e7ef93eae654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.153434 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf" (OuterVolumeSpecName: "kube-api-access-8qvpf") pod "56410990-174e-4fd5-94e7-e7ef93eae654" (UID: "56410990-174e-4fd5-94e7-e7ef93eae654"). InnerVolumeSpecName "kube-api-access-8qvpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.167467 4837 scope.go:117] "RemoveContainer" containerID="f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.215081 4837 scope.go:117] "RemoveContainer" containerID="5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57" Oct 01 08:55:07 crc kubenswrapper[4837]: E1001 08:55:07.218595 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57\": container with ID starting with 5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57 not found: ID does not exist" containerID="5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.218681 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57"} err="failed to get container status \"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57\": rpc error: code = NotFound desc = could not find container \"5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57\": container with ID starting with 5b974abb88317565f8fc29e52ad8681d5ee7463f887f893dd3e074731017ce57 not found: ID does not exist" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.218836 4837 scope.go:117] "RemoveContainer" containerID="4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea" Oct 01 08:55:07 crc kubenswrapper[4837]: E1001 08:55:07.219216 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea\": container with ID starting with 4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea not found: ID does not exist" containerID="4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.219258 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea"} err="failed to get container status \"4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea\": rpc error: code = NotFound desc = could not find container \"4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea\": container with ID starting with 4c9a595b9d1e466135947e3684d616254263ceea58f9deef8ae338e10fb92eea not found: ID does not exist" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.219284 4837 scope.go:117] "RemoveContainer" containerID="f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd" Oct 01 08:55:07 crc kubenswrapper[4837]: E1001 08:55:07.219634 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd\": container with ID starting with f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd not found: ID does not exist" containerID="f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.219668 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd"} err="failed to get container status \"f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd\": rpc error: code = NotFound desc = could not find container \"f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd\": container with ID starting with f92a4450641344693a31d5ec51b1ad99b3440a2cf4e90a693734a9ababef5ddd not found: ID does not exist" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.224007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56410990-174e-4fd5-94e7-e7ef93eae654" (UID: "56410990-174e-4fd5-94e7-e7ef93eae654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.245739 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.245773 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qvpf\" (UniqueName: \"kubernetes.io/projected/56410990-174e-4fd5-94e7-e7ef93eae654-kube-api-access-8qvpf\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.245784 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56410990-174e-4fd5-94e7-e7ef93eae654-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.443528 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.451787 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vl7vv"] Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.832874 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" path="/var/lib/kubelet/pods/56410990-174e-4fd5-94e7-e7ef93eae654/volumes" Oct 01 08:55:07 crc kubenswrapper[4837]: I1001 08:55:07.834902 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" path="/var/lib/kubelet/pods/a65db51c-f0f3-48e6-908f-8b34321296b1/volumes" Oct 01 08:55:09 crc kubenswrapper[4837]: I1001 08:55:09.816005 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:55:09 crc kubenswrapper[4837]: E1001 08:55:09.816761 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:55:21 crc kubenswrapper[4837]: I1001 08:55:21.816514 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:55:21 crc kubenswrapper[4837]: E1001 08:55:21.817373 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:55:34 crc kubenswrapper[4837]: I1001 08:55:34.816856 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:55:34 crc kubenswrapper[4837]: E1001 08:55:34.817829 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:55:48 crc kubenswrapper[4837]: I1001 08:55:48.816844 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:55:48 crc kubenswrapper[4837]: E1001 08:55:48.817869 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:56:03 crc kubenswrapper[4837]: I1001 08:56:03.823937 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:56:03 crc kubenswrapper[4837]: E1001 08:56:03.824745 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.047445 4837 scope.go:117] "RemoveContainer" containerID="f7f2246bfd375936eb2c2b4bd287dfb8c62e101d4b835d7b07d0eb77b18d8d8a" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.074132 4837 scope.go:117] "RemoveContainer" containerID="f62599eb7b29368b067c586f5143bcffe3072109bc10d7d76e1452c533489606" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.109735 4837 scope.go:117] "RemoveContainer" containerID="28a44c948369728e9fa4e1a95e8fc5715453e79653300fb842adcced46acc7b0" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.150002 4837 scope.go:117] "RemoveContainer" containerID="120985a71b11d26ddedf5a3e596de0b8c5e907937ccb4965ccbf0e1705f6609b" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.179396 4837 scope.go:117] "RemoveContainer" containerID="abc94b8067b10a8c31b17162c4a758a9edf1757d265a15aeb2eaa2d86e703387" Oct 01 08:56:06 crc kubenswrapper[4837]: I1001 08:56:06.210809 4837 scope.go:117] "RemoveContainer" containerID="831293fe9fd86d206b4e31dd5d26f9afb31f57575c205c6cf44e4067133dece5" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.116256 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117408 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="extract-utilities" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117428 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="extract-utilities" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117452 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117460 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117472 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="extract-content" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117480 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="extract-content" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117490 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="extract-utilities" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117497 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="extract-utilities" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117518 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="extract-content" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117525 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="extract-content" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.117539 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117546 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117751 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a65db51c-f0f3-48e6-908f-8b34321296b1" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.117775 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="56410990-174e-4fd5-94e7-e7ef93eae654" containerName="registry-server" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.119194 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.135612 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.285289 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.285976 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.286030 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slnst\" (UniqueName: \"kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.387580 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.387717 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.387762 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slnst\" (UniqueName: \"kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.388224 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.388341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.414977 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slnst\" (UniqueName: \"kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst\") pod \"redhat-marketplace-f2ljn\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.460735 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.817185 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:56:18 crc kubenswrapper[4837]: E1001 08:56:18.817589 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 08:56:18 crc kubenswrapper[4837]: I1001 08:56:18.893730 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:19 crc kubenswrapper[4837]: I1001 08:56:19.862218 4837 generic.go:334] "Generic (PLEG): container finished" podID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerID="5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719" exitCode=0 Oct 01 08:56:19 crc kubenswrapper[4837]: I1001 08:56:19.862465 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerDied","Data":"5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719"} Oct 01 08:56:19 crc kubenswrapper[4837]: I1001 08:56:19.862626 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerStarted","Data":"1abfec1ae41b09a9c980107a72622be9933d8a3d88933eb50bcb8b2e205e3210"} Oct 01 08:56:20 crc kubenswrapper[4837]: I1001 08:56:20.874488 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerStarted","Data":"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4"} Oct 01 08:56:21 crc kubenswrapper[4837]: I1001 08:56:21.898417 4837 generic.go:334] "Generic (PLEG): container finished" podID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerID="804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4" exitCode=0 Oct 01 08:56:21 crc kubenswrapper[4837]: I1001 08:56:21.899029 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerDied","Data":"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4"} Oct 01 08:56:21 crc kubenswrapper[4837]: I1001 08:56:21.899109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerStarted","Data":"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10"} Oct 01 08:56:21 crc kubenswrapper[4837]: I1001 08:56:21.924764 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f2ljn" podStartSLOduration=2.150003065 podStartE2EDuration="3.924740002s" podCreationTimestamp="2025-10-01 08:56:18 +0000 UTC" firstStartedPulling="2025-10-01 08:56:19.867397363 +0000 UTC m=+6636.709004838" lastFinishedPulling="2025-10-01 08:56:21.64213432 +0000 UTC m=+6638.483741775" observedRunningTime="2025-10-01 08:56:21.919981685 +0000 UTC m=+6638.761589150" watchObservedRunningTime="2025-10-01 08:56:21.924740002 +0000 UTC m=+6638.766347457" Oct 01 08:56:28 crc kubenswrapper[4837]: I1001 08:56:28.461885 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:28 crc kubenswrapper[4837]: I1001 08:56:28.463013 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:28 crc kubenswrapper[4837]: I1001 08:56:28.518078 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:29 crc kubenswrapper[4837]: I1001 08:56:29.003389 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:29 crc kubenswrapper[4837]: I1001 08:56:29.058667 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:30 crc kubenswrapper[4837]: I1001 08:56:30.978290 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f2ljn" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="registry-server" containerID="cri-o://c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10" gracePeriod=2 Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.455852 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.547076 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities\") pod \"138feef9-ecae-431e-8bc1-edd65e58fa0e\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.547222 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content\") pod \"138feef9-ecae-431e-8bc1-edd65e58fa0e\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.547275 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slnst\" (UniqueName: \"kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst\") pod \"138feef9-ecae-431e-8bc1-edd65e58fa0e\" (UID: \"138feef9-ecae-431e-8bc1-edd65e58fa0e\") " Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.548353 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities" (OuterVolumeSpecName: "utilities") pod "138feef9-ecae-431e-8bc1-edd65e58fa0e" (UID: "138feef9-ecae-431e-8bc1-edd65e58fa0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.556634 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst" (OuterVolumeSpecName: "kube-api-access-slnst") pod "138feef9-ecae-431e-8bc1-edd65e58fa0e" (UID: "138feef9-ecae-431e-8bc1-edd65e58fa0e"). InnerVolumeSpecName "kube-api-access-slnst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.566022 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "138feef9-ecae-431e-8bc1-edd65e58fa0e" (UID: "138feef9-ecae-431e-8bc1-edd65e58fa0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.649575 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.649626 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138feef9-ecae-431e-8bc1-edd65e58fa0e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.649641 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slnst\" (UniqueName: \"kubernetes.io/projected/138feef9-ecae-431e-8bc1-edd65e58fa0e-kube-api-access-slnst\") on node \"crc\" DevicePath \"\"" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.991178 4837 generic.go:334] "Generic (PLEG): container finished" podID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerID="c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10" exitCode=0 Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.991237 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerDied","Data":"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10"} Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.991326 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f2ljn" event={"ID":"138feef9-ecae-431e-8bc1-edd65e58fa0e","Type":"ContainerDied","Data":"1abfec1ae41b09a9c980107a72622be9933d8a3d88933eb50bcb8b2e205e3210"} Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.991352 4837 scope.go:117] "RemoveContainer" containerID="c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10" Oct 01 08:56:31 crc kubenswrapper[4837]: I1001 08:56:31.991260 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f2ljn" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.027132 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.031134 4837 scope.go:117] "RemoveContainer" containerID="804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.036570 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f2ljn"] Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.058395 4837 scope.go:117] "RemoveContainer" containerID="5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.104408 4837 scope.go:117] "RemoveContainer" containerID="c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10" Oct 01 08:56:32 crc kubenswrapper[4837]: E1001 08:56:32.105355 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10\": container with ID starting with c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10 not found: ID does not exist" containerID="c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.105432 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10"} err="failed to get container status \"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10\": rpc error: code = NotFound desc = could not find container \"c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10\": container with ID starting with c016318e1fb7f1b6845e527342e4e1a1f6023e615ab9433d1a842b9fbdd65f10 not found: ID does not exist" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.105472 4837 scope.go:117] "RemoveContainer" containerID="804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4" Oct 01 08:56:32 crc kubenswrapper[4837]: E1001 08:56:32.105997 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4\": container with ID starting with 804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4 not found: ID does not exist" containerID="804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.106049 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4"} err="failed to get container status \"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4\": rpc error: code = NotFound desc = could not find container \"804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4\": container with ID starting with 804b4203a87edac1f46d8aec57407e76c539326b76de4758cd1b270970a98ef4 not found: ID does not exist" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.106089 4837 scope.go:117] "RemoveContainer" containerID="5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719" Oct 01 08:56:32 crc kubenswrapper[4837]: E1001 08:56:32.106885 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719\": container with ID starting with 5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719 not found: ID does not exist" containerID="5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.106928 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719"} err="failed to get container status \"5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719\": rpc error: code = NotFound desc = could not find container \"5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719\": container with ID starting with 5fc12b564c4f87255562651ec3e4de62fd6487cf4154d2e2bdaa1e3b36b4d719 not found: ID does not exist" Oct 01 08:56:32 crc kubenswrapper[4837]: I1001 08:56:32.815327 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 08:56:33 crc kubenswrapper[4837]: I1001 08:56:33.825043 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" path="/var/lib/kubelet/pods/138feef9-ecae-431e-8bc1-edd65e58fa0e/volumes" Oct 01 08:56:34 crc kubenswrapper[4837]: I1001 08:56:34.012524 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5"} Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.765460 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 08:56:35 crc kubenswrapper[4837]: E1001 08:56:35.767437 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="extract-utilities" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.767523 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="extract-utilities" Oct 01 08:56:35 crc kubenswrapper[4837]: E1001 08:56:35.767583 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="extract-content" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.767634 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="extract-content" Oct 01 08:56:35 crc kubenswrapper[4837]: E1001 08:56:35.767716 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="registry-server" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.767783 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="registry-server" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.767984 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="138feef9-ecae-431e-8bc1-edd65e58fa0e" containerName="registry-server" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.768562 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.773330 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cz4sn" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.781850 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.922837 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:35 crc kubenswrapper[4837]: I1001 08:56:35.922954 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5nr9\" (UniqueName: \"kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.024697 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.025339 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5nr9\" (UniqueName: \"kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.029617 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.029651 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1d1729fc028a3a165099c3539a130004f378ba08a7e7c043a6e0cd44dc5e116e/globalmount\"" pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.050031 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5nr9\" (UniqueName: \"kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.069240 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") pod \"mariadb-copy-data\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.100639 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 01 08:56:36 crc kubenswrapper[4837]: I1001 08:56:36.638003 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 08:56:37 crc kubenswrapper[4837]: I1001 08:56:37.035714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"22907b50-b230-4f94-bd39-7d66e4601e08","Type":"ContainerStarted","Data":"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86"} Oct 01 08:56:37 crc kubenswrapper[4837]: I1001 08:56:37.035758 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"22907b50-b230-4f94-bd39-7d66e4601e08","Type":"ContainerStarted","Data":"516954e5bb8b2779be8ed5a3a445596ea05f51ecf5164beab550a5ec221225a9"} Oct 01 08:56:37 crc kubenswrapper[4837]: I1001 08:56:37.056003 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.055980977 podStartE2EDuration="3.055980977s" podCreationTimestamp="2025-10-01 08:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:56:37.052197165 +0000 UTC m=+6653.893804630" watchObservedRunningTime="2025-10-01 08:56:37.055980977 +0000 UTC m=+6653.897588422" Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.372839 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.377216 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.382066 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.484549 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq6j4\" (UniqueName: \"kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4\") pod \"mariadb-client\" (UID: \"e782e35a-864e-41ee-af8c-445ac8ba8f13\") " pod="openstack/mariadb-client" Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.585861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq6j4\" (UniqueName: \"kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4\") pod \"mariadb-client\" (UID: \"e782e35a-864e-41ee-af8c-445ac8ba8f13\") " pod="openstack/mariadb-client" Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.617555 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq6j4\" (UniqueName: \"kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4\") pod \"mariadb-client\" (UID: \"e782e35a-864e-41ee-af8c-445ac8ba8f13\") " pod="openstack/mariadb-client" Oct 01 08:56:39 crc kubenswrapper[4837]: I1001 08:56:39.714001 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:40 crc kubenswrapper[4837]: I1001 08:56:40.199541 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:40 crc kubenswrapper[4837]: W1001 08:56:40.202254 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode782e35a_864e_41ee_af8c_445ac8ba8f13.slice/crio-b26a79632754123a0490afa8d2d160f505c8ba584528876f37b6fb359684a170 WatchSource:0}: Error finding container b26a79632754123a0490afa8d2d160f505c8ba584528876f37b6fb359684a170: Status 404 returned error can't find the container with id b26a79632754123a0490afa8d2d160f505c8ba584528876f37b6fb359684a170 Oct 01 08:56:41 crc kubenswrapper[4837]: I1001 08:56:41.076096 4837 generic.go:334] "Generic (PLEG): container finished" podID="e782e35a-864e-41ee-af8c-445ac8ba8f13" containerID="980788b559a08bbfe5b603cd77525c34f1e8c9892a3a0ed4b54a305dea3dd012" exitCode=0 Oct 01 08:56:41 crc kubenswrapper[4837]: I1001 08:56:41.076645 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e782e35a-864e-41ee-af8c-445ac8ba8f13","Type":"ContainerDied","Data":"980788b559a08bbfe5b603cd77525c34f1e8c9892a3a0ed4b54a305dea3dd012"} Oct 01 08:56:41 crc kubenswrapper[4837]: I1001 08:56:41.076785 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e782e35a-864e-41ee-af8c-445ac8ba8f13","Type":"ContainerStarted","Data":"b26a79632754123a0490afa8d2d160f505c8ba584528876f37b6fb359684a170"} Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.472461 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.496488 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_e782e35a-864e-41ee-af8c-445ac8ba8f13/mariadb-client/0.log" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.521466 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.527439 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.561983 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq6j4\" (UniqueName: \"kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4\") pod \"e782e35a-864e-41ee-af8c-445ac8ba8f13\" (UID: \"e782e35a-864e-41ee-af8c-445ac8ba8f13\") " Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.567570 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4" (OuterVolumeSpecName: "kube-api-access-cq6j4") pod "e782e35a-864e-41ee-af8c-445ac8ba8f13" (UID: "e782e35a-864e-41ee-af8c-445ac8ba8f13"). InnerVolumeSpecName "kube-api-access-cq6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.654491 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:42 crc kubenswrapper[4837]: E1001 08:56:42.654906 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e782e35a-864e-41ee-af8c-445ac8ba8f13" containerName="mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.654931 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e782e35a-864e-41ee-af8c-445ac8ba8f13" containerName="mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.655139 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e782e35a-864e-41ee-af8c-445ac8ba8f13" containerName="mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.655797 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.665966 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq6j4\" (UniqueName: \"kubernetes.io/projected/e782e35a-864e-41ee-af8c-445ac8ba8f13-kube-api-access-cq6j4\") on node \"crc\" DevicePath \"\"" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.671075 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.767844 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xxbt\" (UniqueName: \"kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt\") pod \"mariadb-client\" (UID: \"474256f8-2c7c-423f-a99c-39799022044d\") " pod="openstack/mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.869100 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xxbt\" (UniqueName: \"kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt\") pod \"mariadb-client\" (UID: \"474256f8-2c7c-423f-a99c-39799022044d\") " pod="openstack/mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.888186 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xxbt\" (UniqueName: \"kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt\") pod \"mariadb-client\" (UID: \"474256f8-2c7c-423f-a99c-39799022044d\") " pod="openstack/mariadb-client" Oct 01 08:56:42 crc kubenswrapper[4837]: I1001 08:56:42.987450 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:43 crc kubenswrapper[4837]: I1001 08:56:43.097799 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b26a79632754123a0490afa8d2d160f505c8ba584528876f37b6fb359684a170" Oct 01 08:56:43 crc kubenswrapper[4837]: I1001 08:56:43.097885 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:43 crc kubenswrapper[4837]: I1001 08:56:43.115321 4837 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="e782e35a-864e-41ee-af8c-445ac8ba8f13" podUID="474256f8-2c7c-423f-a99c-39799022044d" Oct 01 08:56:43 crc kubenswrapper[4837]: I1001 08:56:43.416906 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:43 crc kubenswrapper[4837]: W1001 08:56:43.424613 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod474256f8_2c7c_423f_a99c_39799022044d.slice/crio-4b45fc6209e3b7253b1f8048dc0573899d80bdd3dfac7d794fb3751f1b869bc3 WatchSource:0}: Error finding container 4b45fc6209e3b7253b1f8048dc0573899d80bdd3dfac7d794fb3751f1b869bc3: Status 404 returned error can't find the container with id 4b45fc6209e3b7253b1f8048dc0573899d80bdd3dfac7d794fb3751f1b869bc3 Oct 01 08:56:43 crc kubenswrapper[4837]: I1001 08:56:43.826823 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e782e35a-864e-41ee-af8c-445ac8ba8f13" path="/var/lib/kubelet/pods/e782e35a-864e-41ee-af8c-445ac8ba8f13/volumes" Oct 01 08:56:44 crc kubenswrapper[4837]: I1001 08:56:44.128619 4837 generic.go:334] "Generic (PLEG): container finished" podID="474256f8-2c7c-423f-a99c-39799022044d" containerID="e3ba2ee65accb3da5c716ee584ed06166ed5d8a7d2a8015b729b729ded7d8804" exitCode=0 Oct 01 08:56:44 crc kubenswrapper[4837]: I1001 08:56:44.129204 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"474256f8-2c7c-423f-a99c-39799022044d","Type":"ContainerDied","Data":"e3ba2ee65accb3da5c716ee584ed06166ed5d8a7d2a8015b729b729ded7d8804"} Oct 01 08:56:44 crc kubenswrapper[4837]: I1001 08:56:44.129263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"474256f8-2c7c-423f-a99c-39799022044d","Type":"ContainerStarted","Data":"4b45fc6209e3b7253b1f8048dc0573899d80bdd3dfac7d794fb3751f1b869bc3"} Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.463360 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.483286 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_474256f8-2c7c-423f-a99c-39799022044d/mariadb-client/0.log" Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.507848 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.514628 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.616987 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xxbt\" (UniqueName: \"kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt\") pod \"474256f8-2c7c-423f-a99c-39799022044d\" (UID: \"474256f8-2c7c-423f-a99c-39799022044d\") " Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.629741 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt" (OuterVolumeSpecName: "kube-api-access-6xxbt") pod "474256f8-2c7c-423f-a99c-39799022044d" (UID: "474256f8-2c7c-423f-a99c-39799022044d"). InnerVolumeSpecName "kube-api-access-6xxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.719925 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xxbt\" (UniqueName: \"kubernetes.io/projected/474256f8-2c7c-423f-a99c-39799022044d-kube-api-access-6xxbt\") on node \"crc\" DevicePath \"\"" Oct 01 08:56:45 crc kubenswrapper[4837]: I1001 08:56:45.829925 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="474256f8-2c7c-423f-a99c-39799022044d" path="/var/lib/kubelet/pods/474256f8-2c7c-423f-a99c-39799022044d/volumes" Oct 01 08:56:46 crc kubenswrapper[4837]: I1001 08:56:46.146034 4837 scope.go:117] "RemoveContainer" containerID="e3ba2ee65accb3da5c716ee584ed06166ed5d8a7d2a8015b729b729ded7d8804" Oct 01 08:56:46 crc kubenswrapper[4837]: I1001 08:56:46.146349 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.173222 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 08:57:33 crc kubenswrapper[4837]: E1001 08:57:33.174742 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="474256f8-2c7c-423f-a99c-39799022044d" containerName="mariadb-client" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.174777 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="474256f8-2c7c-423f-a99c-39799022044d" containerName="mariadb-client" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.175174 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="474256f8-2c7c-423f-a99c-39799022044d" containerName="mariadb-client" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.176803 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.182959 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.183447 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-kkmvv" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.184655 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.187590 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.188111 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.195379 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.198885 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.213712 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.216348 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.242890 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.253281 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.278800 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.295609 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.295662 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.295724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.295747 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.295956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296133 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-config\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296228 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjh4\" (UniqueName: \"kubernetes.io/projected/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-kube-api-access-4hjh4\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296359 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296413 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296469 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296499 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296519 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-config\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296548 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swrj4\" (UniqueName: \"kubernetes.io/projected/fde602a4-cc23-4e59-bb85-c3923a628294-kube-api-access-swrj4\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296613 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296678 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296739 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bspn\" (UniqueName: \"kubernetes.io/projected/75755bf0-420a-42fd-8d67-0df40028d9ed-kube-api-access-9bspn\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296818 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bd194e9c-8616-4a64-b78f-6142408db747\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bd194e9c-8616-4a64-b78f-6142408db747\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296875 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-config\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296903 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296940 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296964 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.296988 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398258 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398298 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398323 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-config\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398349 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398375 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjh4\" (UniqueName: \"kubernetes.io/projected/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-kube-api-access-4hjh4\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398411 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398436 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398462 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-config\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398508 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398536 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swrj4\" (UniqueName: \"kubernetes.io/projected/fde602a4-cc23-4e59-bb85-c3923a628294-kube-api-access-swrj4\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398568 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398596 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398622 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bspn\" (UniqueName: \"kubernetes.io/projected/75755bf0-420a-42fd-8d67-0df40028d9ed-kube-api-access-9bspn\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398661 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bd194e9c-8616-4a64-b78f-6142408db747\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bd194e9c-8616-4a64-b78f-6142408db747\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398683 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398733 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-config\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398760 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398790 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398812 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.398838 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.399074 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.399300 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.400094 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-config\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.400152 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fde602a4-cc23-4e59-bb85-c3923a628294-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.400261 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.401180 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.401537 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-config\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.401660 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.404845 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75755bf0-420a-42fd-8d67-0df40028d9ed-config\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.408277 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.413202 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.415574 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.415738 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.415780 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/47c3e807b8b7aa8382eb0c118cabfae92ae4cff730b5fd2af46a68c7cb8c64ed/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.415954 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bd194e9c-8616-4a64-b78f-6142408db747\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bd194e9c-8616-4a64-b78f-6142408db747\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4ba3dcd7e059bd731b9d1a18152b10c6111aae150186f018899f5f83c963f6f1/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.417737 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.419027 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.419061 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.420048 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.420079 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afb24ccc5b85f64cb66e483f10c6922b9acc4a1a6ceb4547fd6a0c3bd6e7e282/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.420071 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.420461 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fde602a4-cc23-4e59-bb85-c3923a628294-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.421011 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75755bf0-420a-42fd-8d67-0df40028d9ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.421785 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.423101 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swrj4\" (UniqueName: \"kubernetes.io/projected/fde602a4-cc23-4e59-bb85-c3923a628294-kube-api-access-swrj4\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.423356 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bspn\" (UniqueName: \"kubernetes.io/projected/75755bf0-420a-42fd-8d67-0df40028d9ed-kube-api-access-9bspn\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.434602 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjh4\" (UniqueName: \"kubernetes.io/projected/ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b-kube-api-access-4hjh4\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.448312 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cacc95b4-17cf-4657-b635-6b86ba94c08f\") pod \"ovsdbserver-nb-0\" (UID: \"fde602a4-cc23-4e59-bb85-c3923a628294\") " pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.451491 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cc341b-3fd3-44fb-ab92-921cc2440724\") pod \"ovsdbserver-nb-1\" (UID: \"75755bf0-420a-42fd-8d67-0df40028d9ed\") " pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.452025 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bd194e9c-8616-4a64-b78f-6142408db747\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bd194e9c-8616-4a64-b78f-6142408db747\") pod \"ovsdbserver-nb-2\" (UID: \"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b\") " pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.535954 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.555478 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:33 crc kubenswrapper[4837]: I1001 08:57:33.569953 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.206614 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.306209 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 01 08:57:34 crc kubenswrapper[4837]: W1001 08:57:34.306813 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca1a22b9_7313_4f9f_a9f4_b2efe0aa508b.slice/crio-da915c59c4ea7ee8d588201d1c761a3a068b67bf0a57337c16b16768fed0a853 WatchSource:0}: Error finding container da915c59c4ea7ee8d588201d1c761a3a068b67bf0a57337c16b16768fed0a853: Status 404 returned error can't find the container with id da915c59c4ea7ee8d588201d1c761a3a068b67bf0a57337c16b16768fed0a853 Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.566640 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b","Type":"ContainerStarted","Data":"da915c59c4ea7ee8d588201d1c761a3a068b67bf0a57337c16b16768fed0a853"} Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.568030 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fde602a4-cc23-4e59-bb85-c3923a628294","Type":"ContainerStarted","Data":"4e9c9cad95b548fe7e91021c7ff700c1f70e6c60f9da33c04ddb39d9f766911f"} Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.920827 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.923614 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.929142 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.929174 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-z99cq" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.929265 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.929300 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.935212 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.953795 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.955569 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.973387 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.974936 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:34 crc kubenswrapper[4837]: I1001 08:57:34.988613 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.009194 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037026 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b867c980-9d79-4532-8010-d8b24f147eba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b867c980-9d79-4532-8010-d8b24f147eba\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037122 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-config\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037171 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037201 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtg2\" (UniqueName: \"kubernetes.io/projected/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-kube-api-access-fwtg2\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.037254 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139133 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139190 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrlbl\" (UniqueName: \"kubernetes.io/projected/679a277e-9e24-4a5a-ad48-6ba934d0a55a-kube-api-access-lrlbl\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139247 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139271 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139301 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-config\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139331 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b867c980-9d79-4532-8010-d8b24f147eba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b867c980-9d79-4532-8010-d8b24f147eba\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139355 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139393 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139417 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f6d4b498-3d59-40df-bab2-9940d84db536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6d4b498-3d59-40df-bab2-9940d84db536\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-config\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139468 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139491 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139529 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139562 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139581 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p42hs\" (UniqueName: \"kubernetes.io/projected/680fbe7d-0b4f-4755-b18e-21436e4c0530-kube-api-access-p42hs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139608 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139670 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-config\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139715 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.139740 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtg2\" (UniqueName: \"kubernetes.io/projected/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-kube-api-access-fwtg2\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.141033 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.141064 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.141099 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.141138 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.141169 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.146255 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.147130 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-config\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.148270 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.151066 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.151127 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b867c980-9d79-4532-8010-d8b24f147eba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b867c980-9d79-4532-8010-d8b24f147eba\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a86b905006225677994acb727828e7cb99d264a3a0644142621a1da3b12adb7e/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.151536 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.151566 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.157458 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.169185 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.170815 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtg2\" (UniqueName: \"kubernetes.io/projected/1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416-kube-api-access-fwtg2\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: W1001 08:57:35.173685 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75755bf0_420a_42fd_8d67_0df40028d9ed.slice/crio-e710256a05b60e892b35c50d64a93802c55b0e846f1044beee3e74327bf689ce WatchSource:0}: Error finding container e710256a05b60e892b35c50d64a93802c55b0e846f1044beee3e74327bf689ce: Status 404 returned error can't find the container with id e710256a05b60e892b35c50d64a93802c55b0e846f1044beee3e74327bf689ce Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.203100 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b867c980-9d79-4532-8010-d8b24f147eba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b867c980-9d79-4532-8010-d8b24f147eba\") pod \"ovsdbserver-sb-0\" (UID: \"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416\") " pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.242868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-config\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.242944 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.242977 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243013 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243054 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243109 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243131 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrlbl\" (UniqueName: \"kubernetes.io/projected/679a277e-9e24-4a5a-ad48-6ba934d0a55a-kube-api-access-lrlbl\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243228 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243268 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-config\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243297 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243342 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f6d4b498-3d59-40df-bab2-9940d84db536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6d4b498-3d59-40df-bab2-9940d84db536\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243381 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243440 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243464 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p42hs\" (UniqueName: \"kubernetes.io/projected/680fbe7d-0b4f-4755-b18e-21436e4c0530-kube-api-access-p42hs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.243494 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.245046 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.245464 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.245899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a277e-9e24-4a5a-ad48-6ba934d0a55a-config\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.246139 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-config\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.246455 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.246773 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/680fbe7d-0b4f-4755-b18e-21436e4c0530-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.249952 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.250863 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.250891 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f6d4b498-3d59-40df-bab2-9940d84db536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6d4b498-3d59-40df-bab2-9940d84db536\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b84f2c300fc929b3121cac23a9b06aba82c8a12fcd635dbb4fd410fb62eea370/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.251631 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.251684 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/114a504fb84607c37a7e8fd523a91317626974f27c9f0115eb2ed9229e8b75a1/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.252834 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.253106 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.254344 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.255301 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/680fbe7d-0b4f-4755-b18e-21436e4c0530-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.255848 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.262844 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a277e-9e24-4a5a-ad48-6ba934d0a55a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.269388 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrlbl\" (UniqueName: \"kubernetes.io/projected/679a277e-9e24-4a5a-ad48-6ba934d0a55a-kube-api-access-lrlbl\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.269461 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p42hs\" (UniqueName: \"kubernetes.io/projected/680fbe7d-0b4f-4755-b18e-21436e4c0530-kube-api-access-p42hs\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.313087 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f6d4b498-3d59-40df-bab2-9940d84db536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6d4b498-3d59-40df-bab2-9940d84db536\") pod \"ovsdbserver-sb-1\" (UID: \"679a277e-9e24-4a5a-ad48-6ba934d0a55a\") " pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.315003 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2f56e895-2591-4ae2-9b2f-413d910645c8\") pod \"ovsdbserver-sb-2\" (UID: \"680fbe7d-0b4f-4755-b18e-21436e4c0530\") " pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.577177 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.579977 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"75755bf0-420a-42fd-8d67-0df40028d9ed","Type":"ContainerStarted","Data":"e710256a05b60e892b35c50d64a93802c55b0e846f1044beee3e74327bf689ce"} Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.599638 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:35 crc kubenswrapper[4837]: I1001 08:57:35.856870 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 08:57:35 crc kubenswrapper[4837]: W1001 08:57:35.870265 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fbaf3d0_9cb4_4e03_bd2e_1ab9ea38d416.slice/crio-ca39dfa59309f3598dc5cfa5897e27e8ce5beb37a98796a5a9d41c432e446552 WatchSource:0}: Error finding container ca39dfa59309f3598dc5cfa5897e27e8ce5beb37a98796a5a9d41c432e446552: Status 404 returned error can't find the container with id ca39dfa59309f3598dc5cfa5897e27e8ce5beb37a98796a5a9d41c432e446552 Oct 01 08:57:36 crc kubenswrapper[4837]: I1001 08:57:36.243676 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 01 08:57:36 crc kubenswrapper[4837]: W1001 08:57:36.258416 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod680fbe7d_0b4f_4755_b18e_21436e4c0530.slice/crio-c0f1432e12b592643223e7929b7d0599581ce0cf1f0092c90c83347af7dbdfba WatchSource:0}: Error finding container c0f1432e12b592643223e7929b7d0599581ce0cf1f0092c90c83347af7dbdfba: Status 404 returned error can't find the container with id c0f1432e12b592643223e7929b7d0599581ce0cf1f0092c90c83347af7dbdfba Oct 01 08:57:36 crc kubenswrapper[4837]: I1001 08:57:36.350760 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 01 08:57:36 crc kubenswrapper[4837]: W1001 08:57:36.359885 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod679a277e_9e24_4a5a_ad48_6ba934d0a55a.slice/crio-384496be89ef538eaf2bb3d5f54f22fffd6944f38635c22694ef5d67f54a0572 WatchSource:0}: Error finding container 384496be89ef538eaf2bb3d5f54f22fffd6944f38635c22694ef5d67f54a0572: Status 404 returned error can't find the container with id 384496be89ef538eaf2bb3d5f54f22fffd6944f38635c22694ef5d67f54a0572 Oct 01 08:57:36 crc kubenswrapper[4837]: I1001 08:57:36.601096 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"680fbe7d-0b4f-4755-b18e-21436e4c0530","Type":"ContainerStarted","Data":"c0f1432e12b592643223e7929b7d0599581ce0cf1f0092c90c83347af7dbdfba"} Oct 01 08:57:36 crc kubenswrapper[4837]: I1001 08:57:36.603346 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416","Type":"ContainerStarted","Data":"ca39dfa59309f3598dc5cfa5897e27e8ce5beb37a98796a5a9d41c432e446552"} Oct 01 08:57:36 crc kubenswrapper[4837]: I1001 08:57:36.605027 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"679a277e-9e24-4a5a-ad48-6ba934d0a55a","Type":"ContainerStarted","Data":"384496be89ef538eaf2bb3d5f54f22fffd6944f38635c22694ef5d67f54a0572"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.674344 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fde602a4-cc23-4e59-bb85-c3923a628294","Type":"ContainerStarted","Data":"f6a90ba07e72822b55474adfd9aba884f8dab21cf3bd50c89254a6f5ac696d56"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.675132 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"fde602a4-cc23-4e59-bb85-c3923a628294","Type":"ContainerStarted","Data":"72fe1276aeb382da72328e8371e560b03afcd5225db4476d38488e6755a8a9ee"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.679181 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416","Type":"ContainerStarted","Data":"50d5e974c74fe21f7311a9437292bd9d568eb000764c30d5b3edce973d7b85eb"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.679222 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416","Type":"ContainerStarted","Data":"fd9678e9bb8f5660bf47daf24c690fbe93ee2b693fec3f194f464677996ff33d"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.686761 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"679a277e-9e24-4a5a-ad48-6ba934d0a55a","Type":"ContainerStarted","Data":"1b3477e08e9f2a3effb461f2b7dcffb15c6136880af9eae7bc1038594a39e120"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.686802 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"679a277e-9e24-4a5a-ad48-6ba934d0a55a","Type":"ContainerStarted","Data":"0dd13cd5caef7a1a103404fef4059f59aeb3e6d02519616be45e3e9b1579d923"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.693217 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b","Type":"ContainerStarted","Data":"c599f3de00e54387ccf33fcd456dc4a277ecf016343723913387eb22a9622798"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.709510 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.871549695 podStartE2EDuration="10.709482113s" podCreationTimestamp="2025-10-01 08:57:32 +0000 UTC" firstStartedPulling="2025-10-01 08:57:34.22120989 +0000 UTC m=+6711.062817345" lastFinishedPulling="2025-10-01 08:57:42.059142288 +0000 UTC m=+6718.900749763" observedRunningTime="2025-10-01 08:57:42.69881624 +0000 UTC m=+6719.540423695" watchObservedRunningTime="2025-10-01 08:57:42.709482113 +0000 UTC m=+6719.551089568" Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.718346 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"75755bf0-420a-42fd-8d67-0df40028d9ed","Type":"ContainerStarted","Data":"05a85ed3827ee687fa7978e6c518e687466f253448f6102dc6378d0130661703"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.718393 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"75755bf0-420a-42fd-8d67-0df40028d9ed","Type":"ContainerStarted","Data":"aaccdc459f1c63d37aa4664f733275e33f88d404b5b5315415e61ade0843fe21"} Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.750325 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.288964738 podStartE2EDuration="9.750305825s" podCreationTimestamp="2025-10-01 08:57:33 +0000 UTC" firstStartedPulling="2025-10-01 08:57:36.363899095 +0000 UTC m=+6713.205506550" lastFinishedPulling="2025-10-01 08:57:41.825240182 +0000 UTC m=+6718.666847637" observedRunningTime="2025-10-01 08:57:42.728471169 +0000 UTC m=+6719.570078624" watchObservedRunningTime="2025-10-01 08:57:42.750305825 +0000 UTC m=+6719.591913280" Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.755867 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.8172801720000002 podStartE2EDuration="9.755854792s" podCreationTimestamp="2025-10-01 08:57:33 +0000 UTC" firstStartedPulling="2025-10-01 08:57:35.874275477 +0000 UTC m=+6712.715882952" lastFinishedPulling="2025-10-01 08:57:41.812850117 +0000 UTC m=+6718.654457572" observedRunningTime="2025-10-01 08:57:42.746861021 +0000 UTC m=+6719.588468476" watchObservedRunningTime="2025-10-01 08:57:42.755854792 +0000 UTC m=+6719.597462247" Oct 01 08:57:42 crc kubenswrapper[4837]: I1001 08:57:42.778591 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.144690899 podStartE2EDuration="10.778558529s" podCreationTimestamp="2025-10-01 08:57:32 +0000 UTC" firstStartedPulling="2025-10-01 08:57:35.177024269 +0000 UTC m=+6712.018631724" lastFinishedPulling="2025-10-01 08:57:41.810891899 +0000 UTC m=+6718.652499354" observedRunningTime="2025-10-01 08:57:42.771389533 +0000 UTC m=+6719.612996988" watchObservedRunningTime="2025-10-01 08:57:42.778558529 +0000 UTC m=+6719.620165984" Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.536254 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.570840 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.730420 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"680fbe7d-0b4f-4755-b18e-21436e4c0530","Type":"ContainerStarted","Data":"7d50cbe26ce1123a9823d066c51dae92dd066c79710a48a8377fef9f35f8ac4d"} Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.730497 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"680fbe7d-0b4f-4755-b18e-21436e4c0530","Type":"ContainerStarted","Data":"6378d145a1eebc12022ab5ab56ff45eafd50128fce4b52abb0bd0fc5b78899f3"} Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.734227 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b","Type":"ContainerStarted","Data":"36ccf12a71bc596a311694b555a897ecf24c1839e315ab6dc679561b6f0b2bfc"} Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.770675 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.443577907 podStartE2EDuration="10.77064208s" podCreationTimestamp="2025-10-01 08:57:33 +0000 UTC" firstStartedPulling="2025-10-01 08:57:36.260472134 +0000 UTC m=+6713.102079589" lastFinishedPulling="2025-10-01 08:57:42.587536307 +0000 UTC m=+6719.429143762" observedRunningTime="2025-10-01 08:57:43.760993463 +0000 UTC m=+6720.602600958" watchObservedRunningTime="2025-10-01 08:57:43.77064208 +0000 UTC m=+6720.612249535" Oct 01 08:57:43 crc kubenswrapper[4837]: I1001 08:57:43.798124 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.794884408 podStartE2EDuration="11.798096315s" podCreationTimestamp="2025-10-01 08:57:32 +0000 UTC" firstStartedPulling="2025-10-01 08:57:34.308891164 +0000 UTC m=+6711.150498629" lastFinishedPulling="2025-10-01 08:57:42.312103081 +0000 UTC m=+6719.153710536" observedRunningTime="2025-10-01 08:57:43.796540446 +0000 UTC m=+6720.638147951" watchObservedRunningTime="2025-10-01 08:57:43.798096315 +0000 UTC m=+6720.639703770" Oct 01 08:57:44 crc kubenswrapper[4837]: I1001 08:57:44.256941 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:44 crc kubenswrapper[4837]: I1001 08:57:44.578265 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:44 crc kubenswrapper[4837]: I1001 08:57:44.600784 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.256455 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.536223 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.557073 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.571008 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.577779 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.588503 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.600239 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.630872 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.657572 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:45 crc kubenswrapper[4837]: I1001 08:57:45.752646 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.332407 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.392296 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.641306 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.655474 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.687911 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.689250 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.704449 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.711471 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.755671 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.756022 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.756082 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5zhd\" (UniqueName: \"kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.756130 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.802792 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.859153 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.859211 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.859385 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.862806 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5zhd\" (UniqueName: \"kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.864340 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.865410 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.866278 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.938465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5zhd\" (UniqueName: \"kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd\") pod \"dnsmasq-dns-784dc555dc-2bzj2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:47 crc kubenswrapper[4837]: I1001 08:57:47.965511 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.015177 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.516253 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:48 crc kubenswrapper[4837]: W1001 08:57:48.522241 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf732303_e264_43f3_93ba_7b0cf7696ed2.slice/crio-4ad21cf5be769e8d11d1a739fc68ade404b1831e7fca27f5572dbd0154c0ee89 WatchSource:0}: Error finding container 4ad21cf5be769e8d11d1a739fc68ade404b1831e7fca27f5572dbd0154c0ee89: Status 404 returned error can't find the container with id 4ad21cf5be769e8d11d1a739fc68ade404b1831e7fca27f5572dbd0154c0ee89 Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.603538 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.625337 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.646993 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.793783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" event={"ID":"df732303-e264-43f3-93ba-7b0cf7696ed2","Type":"ContainerStarted","Data":"4ad21cf5be769e8d11d1a739fc68ade404b1831e7fca27f5572dbd0154c0ee89"} Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.830471 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.864460 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.866990 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.870395 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.874875 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.986304 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.986420 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzvg\" (UniqueName: \"kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.986492 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.986536 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:48 crc kubenswrapper[4837]: I1001 08:57:48.986558 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.088097 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.088187 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.088212 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.088261 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.088302 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzvg\" (UniqueName: \"kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.089862 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.089904 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.089868 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.090089 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.112848 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzvg\" (UniqueName: \"kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg\") pod \"dnsmasq-dns-7744d59f9-95kt7\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.186148 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.648302 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:57:49 crc kubenswrapper[4837]: W1001 08:57:49.649002 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a56f34_a1fd_42c8_a634_bae396251457.slice/crio-ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec WatchSource:0}: Error finding container ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec: Status 404 returned error can't find the container with id ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.803275 4837 generic.go:334] "Generic (PLEG): container finished" podID="df732303-e264-43f3-93ba-7b0cf7696ed2" containerID="9e4f0105655927d8385b732f4faa45ffbc4d27a76dfc329ac232cd1b12d4039d" exitCode=0 Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.803391 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" event={"ID":"df732303-e264-43f3-93ba-7b0cf7696ed2","Type":"ContainerDied","Data":"9e4f0105655927d8385b732f4faa45ffbc4d27a76dfc329ac232cd1b12d4039d"} Oct 01 08:57:49 crc kubenswrapper[4837]: I1001 08:57:49.805313 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" event={"ID":"23a56f34-a1fd-42c8-a634-bae396251457","Type":"ContainerStarted","Data":"ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec"} Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.092678 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.208549 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5zhd\" (UniqueName: \"kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd\") pod \"df732303-e264-43f3-93ba-7b0cf7696ed2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.208630 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config\") pod \"df732303-e264-43f3-93ba-7b0cf7696ed2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.208792 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb\") pod \"df732303-e264-43f3-93ba-7b0cf7696ed2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.208896 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc\") pod \"df732303-e264-43f3-93ba-7b0cf7696ed2\" (UID: \"df732303-e264-43f3-93ba-7b0cf7696ed2\") " Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.214921 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd" (OuterVolumeSpecName: "kube-api-access-x5zhd") pod "df732303-e264-43f3-93ba-7b0cf7696ed2" (UID: "df732303-e264-43f3-93ba-7b0cf7696ed2"). InnerVolumeSpecName "kube-api-access-x5zhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.231326 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df732303-e264-43f3-93ba-7b0cf7696ed2" (UID: "df732303-e264-43f3-93ba-7b0cf7696ed2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.232054 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df732303-e264-43f3-93ba-7b0cf7696ed2" (UID: "df732303-e264-43f3-93ba-7b0cf7696ed2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.246047 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config" (OuterVolumeSpecName: "config") pod "df732303-e264-43f3-93ba-7b0cf7696ed2" (UID: "df732303-e264-43f3-93ba-7b0cf7696ed2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.310996 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5zhd\" (UniqueName: \"kubernetes.io/projected/df732303-e264-43f3-93ba-7b0cf7696ed2-kube-api-access-x5zhd\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.311033 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.311044 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.311053 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df732303-e264-43f3-93ba-7b0cf7696ed2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.813857 4837 generic.go:334] "Generic (PLEG): container finished" podID="23a56f34-a1fd-42c8-a634-bae396251457" containerID="d710015d36f33963b29c241e3ce92ee592a58e400662174ab274488ac0229639" exitCode=0 Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.813930 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" event={"ID":"23a56f34-a1fd-42c8-a634-bae396251457","Type":"ContainerDied","Data":"d710015d36f33963b29c241e3ce92ee592a58e400662174ab274488ac0229639"} Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.817832 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" event={"ID":"df732303-e264-43f3-93ba-7b0cf7696ed2","Type":"ContainerDied","Data":"4ad21cf5be769e8d11d1a739fc68ade404b1831e7fca27f5572dbd0154c0ee89"} Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.818170 4837 scope.go:117] "RemoveContainer" containerID="9e4f0105655927d8385b732f4faa45ffbc4d27a76dfc329ac232cd1b12d4039d" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.817924 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784dc555dc-2bzj2" Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.926595 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:50 crc kubenswrapper[4837]: I1001 08:57:50.971324 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784dc555dc-2bzj2"] Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.677777 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 01 08:57:51 crc kubenswrapper[4837]: E1001 08:57:51.678257 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df732303-e264-43f3-93ba-7b0cf7696ed2" containerName="init" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.678285 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="df732303-e264-43f3-93ba-7b0cf7696ed2" containerName="init" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.678491 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="df732303-e264-43f3-93ba-7b0cf7696ed2" containerName="init" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.679363 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.682296 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.687526 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.827445 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df732303-e264-43f3-93ba-7b0cf7696ed2" path="/var/lib/kubelet/pods/df732303-e264-43f3-93ba-7b0cf7696ed2/volumes" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.828014 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" event={"ID":"23a56f34-a1fd-42c8-a634-bae396251457","Type":"ContainerStarted","Data":"2e31be9975305e39a5602ea5a2b1bfee78f1113567b1ccb6803ce3073163ed37"} Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.828047 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.842225 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9hgf\" (UniqueName: \"kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.842544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.842728 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.853421 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" podStartSLOduration=3.853402811 podStartE2EDuration="3.853402811s" podCreationTimestamp="2025-10-01 08:57:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:57:51.850385307 +0000 UTC m=+6728.691992772" watchObservedRunningTime="2025-10-01 08:57:51.853402811 +0000 UTC m=+6728.695010266" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.944549 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.944669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9hgf\" (UniqueName: \"kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.944803 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.948742 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.948797 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bdc5be76aca342ba831580857d186750a6d3bd3c2c69e6c66300c58dadfe6c1e/globalmount\"" pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.951803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.963054 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9hgf\" (UniqueName: \"kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:51 crc kubenswrapper[4837]: I1001 08:57:51.986048 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") pod \"ovn-copy-data\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " pod="openstack/ovn-copy-data" Oct 01 08:57:52 crc kubenswrapper[4837]: I1001 08:57:52.001586 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 01 08:57:52 crc kubenswrapper[4837]: I1001 08:57:52.527331 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 01 08:57:52 crc kubenswrapper[4837]: I1001 08:57:52.839650 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"8e0d4988-7106-406d-aaf1-02b307230818","Type":"ContainerStarted","Data":"675ba22379501f672d50919a77e1b503f0fdfb769db3a3e45b8ba4f4c4372bd7"} Oct 01 08:57:53 crc kubenswrapper[4837]: I1001 08:57:53.865389 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"8e0d4988-7106-406d-aaf1-02b307230818","Type":"ContainerStarted","Data":"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861"} Oct 01 08:57:53 crc kubenswrapper[4837]: I1001 08:57:53.889140 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.346594262 podStartE2EDuration="3.889114988s" podCreationTimestamp="2025-10-01 08:57:50 +0000 UTC" firstStartedPulling="2025-10-01 08:57:52.542419507 +0000 UTC m=+6729.384026982" lastFinishedPulling="2025-10-01 08:57:53.084940253 +0000 UTC m=+6729.926547708" observedRunningTime="2025-10-01 08:57:53.88633487 +0000 UTC m=+6730.727942335" watchObservedRunningTime="2025-10-01 08:57:53.889114988 +0000 UTC m=+6730.730722453" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.188191 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.263099 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.263537 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="dnsmasq-dns" containerID="cri-o://5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00" gracePeriod=10 Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.740851 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.817959 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc\") pod \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.818039 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config\") pod \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.818095 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62x6\" (UniqueName: \"kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6\") pod \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\" (UID: \"9275d07b-d7b2-4d5e-bee1-296a9115b49a\") " Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.826201 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6" (OuterVolumeSpecName: "kube-api-access-m62x6") pod "9275d07b-d7b2-4d5e-bee1-296a9115b49a" (UID: "9275d07b-d7b2-4d5e-bee1-296a9115b49a"). InnerVolumeSpecName "kube-api-access-m62x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.870784 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config" (OuterVolumeSpecName: "config") pod "9275d07b-d7b2-4d5e-bee1-296a9115b49a" (UID: "9275d07b-d7b2-4d5e-bee1-296a9115b49a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.875347 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9275d07b-d7b2-4d5e-bee1-296a9115b49a" (UID: "9275d07b-d7b2-4d5e-bee1-296a9115b49a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.920222 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.920270 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9275d07b-d7b2-4d5e-bee1-296a9115b49a-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.920289 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62x6\" (UniqueName: \"kubernetes.io/projected/9275d07b-d7b2-4d5e-bee1-296a9115b49a-kube-api-access-m62x6\") on node \"crc\" DevicePath \"\"" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.923005 4837 generic.go:334] "Generic (PLEG): container finished" podID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerID="5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00" exitCode=0 Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.923055 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.923055 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" event={"ID":"9275d07b-d7b2-4d5e-bee1-296a9115b49a","Type":"ContainerDied","Data":"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00"} Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.923344 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84966d4785-xjgcs" event={"ID":"9275d07b-d7b2-4d5e-bee1-296a9115b49a","Type":"ContainerDied","Data":"0de9561a6bec652e89f65f83b71088fd894d8b4c465d012595040d576282705c"} Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.923426 4837 scope.go:117] "RemoveContainer" containerID="5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.954400 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.955370 4837 scope.go:117] "RemoveContainer" containerID="5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.961072 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84966d4785-xjgcs"] Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.991373 4837 scope.go:117] "RemoveContainer" containerID="5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00" Oct 01 08:57:59 crc kubenswrapper[4837]: E1001 08:57:59.991989 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00\": container with ID starting with 5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00 not found: ID does not exist" containerID="5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.992031 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00"} err="failed to get container status \"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00\": rpc error: code = NotFound desc = could not find container \"5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00\": container with ID starting with 5afd3ebd361ba8990fa9ea6492e85a5a41d339ceea3e3467c89af463df344c00 not found: ID does not exist" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.992060 4837 scope.go:117] "RemoveContainer" containerID="5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6" Oct 01 08:57:59 crc kubenswrapper[4837]: E1001 08:57:59.992495 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6\": container with ID starting with 5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6 not found: ID does not exist" containerID="5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6" Oct 01 08:57:59 crc kubenswrapper[4837]: I1001 08:57:59.992519 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6"} err="failed to get container status \"5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6\": rpc error: code = NotFound desc = could not find container \"5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6\": container with ID starting with 5be4cdc2b5bdaacb0ad1b8fa1d07a6b7fd908040be336eb0b7ca210acb2ed0f6 not found: ID does not exist" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.707927 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 01 08:58:01 crc kubenswrapper[4837]: E1001 08:58:01.708432 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="dnsmasq-dns" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.708454 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="dnsmasq-dns" Oct 01 08:58:01 crc kubenswrapper[4837]: E1001 08:58:01.708475 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="init" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.708482 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="init" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.708680 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" containerName="dnsmasq-dns" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.709903 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.711912 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.712169 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.715059 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.716402 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5qt5z" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.727817 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.757873 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-config\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758337 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758371 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2h7z\" (UniqueName: \"kubernetes.io/projected/de8d83c2-5603-4289-bff7-5121a228c96b-kube-api-access-s2h7z\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758502 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-scripts\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758576 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.758602 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.827190 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9275d07b-d7b2-4d5e-bee1-296a9115b49a" path="/var/lib/kubelet/pods/9275d07b-d7b2-4d5e-bee1-296a9115b49a/volumes" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.860566 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-config\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.860632 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.860736 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.860859 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2h7z\" (UniqueName: \"kubernetes.io/projected/de8d83c2-5603-4289-bff7-5121a228c96b-kube-api-access-s2h7z\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.860938 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-scripts\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.861029 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.861074 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.862288 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-config\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.862539 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.862999 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de8d83c2-5603-4289-bff7-5121a228c96b-scripts\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.866165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.867667 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.868490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8d83c2-5603-4289-bff7-5121a228c96b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:01 crc kubenswrapper[4837]: I1001 08:58:01.885627 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2h7z\" (UniqueName: \"kubernetes.io/projected/de8d83c2-5603-4289-bff7-5121a228c96b-kube-api-access-s2h7z\") pod \"ovn-northd-0\" (UID: \"de8d83c2-5603-4289-bff7-5121a228c96b\") " pod="openstack/ovn-northd-0" Oct 01 08:58:02 crc kubenswrapper[4837]: I1001 08:58:02.035105 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 08:58:02 crc kubenswrapper[4837]: I1001 08:58:02.315082 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 08:58:02 crc kubenswrapper[4837]: I1001 08:58:02.952961 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"de8d83c2-5603-4289-bff7-5121a228c96b","Type":"ContainerStarted","Data":"ba46a316496948db3009a71062c20654a6de3a3267d3af5582c1b9f32aa3fe92"} Oct 01 08:58:03 crc kubenswrapper[4837]: I1001 08:58:03.964822 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"de8d83c2-5603-4289-bff7-5121a228c96b","Type":"ContainerStarted","Data":"86a9d87b82a99e50674e90490ef41ef088fabce7b2176030b686a72b48484660"} Oct 01 08:58:03 crc kubenswrapper[4837]: I1001 08:58:03.965212 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"de8d83c2-5603-4289-bff7-5121a228c96b","Type":"ContainerStarted","Data":"3653f20252a74fe126cc686a7cb6fee2749fefd99a8b129331fcae22ccf85b21"} Oct 01 08:58:03 crc kubenswrapper[4837]: I1001 08:58:03.967077 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 01 08:58:04 crc kubenswrapper[4837]: I1001 08:58:04.007277 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.177636618 podStartE2EDuration="3.007225997s" podCreationTimestamp="2025-10-01 08:58:01 +0000 UTC" firstStartedPulling="2025-10-01 08:58:02.328230983 +0000 UTC m=+6739.169838438" lastFinishedPulling="2025-10-01 08:58:03.157820352 +0000 UTC m=+6739.999427817" observedRunningTime="2025-10-01 08:58:03.994516465 +0000 UTC m=+6740.836123980" watchObservedRunningTime="2025-10-01 08:58:04.007225997 +0000 UTC m=+6740.848833492" Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.403795 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g5l26"] Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.405496 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.414594 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g5l26"] Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.544010 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pvdp\" (UniqueName: \"kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp\") pod \"keystone-db-create-g5l26\" (UID: \"71a25257-e61a-493c-ae6b-42a29d97a6cd\") " pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.645961 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pvdp\" (UniqueName: \"kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp\") pod \"keystone-db-create-g5l26\" (UID: \"71a25257-e61a-493c-ae6b-42a29d97a6cd\") " pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.669569 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pvdp\" (UniqueName: \"kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp\") pod \"keystone-db-create-g5l26\" (UID: \"71a25257-e61a-493c-ae6b-42a29d97a6cd\") " pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:09 crc kubenswrapper[4837]: I1001 08:58:09.725847 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:10 crc kubenswrapper[4837]: I1001 08:58:10.233331 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g5l26"] Oct 01 08:58:10 crc kubenswrapper[4837]: W1001 08:58:10.241360 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71a25257_e61a_493c_ae6b_42a29d97a6cd.slice/crio-f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0 WatchSource:0}: Error finding container f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0: Status 404 returned error can't find the container with id f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0 Oct 01 08:58:11 crc kubenswrapper[4837]: I1001 08:58:11.057592 4837 generic.go:334] "Generic (PLEG): container finished" podID="71a25257-e61a-493c-ae6b-42a29d97a6cd" containerID="473552225d6a3ada480afee842621cdd2f341e1545cefea837cb137e985b3558" exitCode=0 Oct 01 08:58:11 crc kubenswrapper[4837]: I1001 08:58:11.057666 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g5l26" event={"ID":"71a25257-e61a-493c-ae6b-42a29d97a6cd","Type":"ContainerDied","Data":"473552225d6a3ada480afee842621cdd2f341e1545cefea837cb137e985b3558"} Oct 01 08:58:11 crc kubenswrapper[4837]: I1001 08:58:11.057731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g5l26" event={"ID":"71a25257-e61a-493c-ae6b-42a29d97a6cd","Type":"ContainerStarted","Data":"f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0"} Oct 01 08:58:12 crc kubenswrapper[4837]: I1001 08:58:12.487629 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:12 crc kubenswrapper[4837]: I1001 08:58:12.598189 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pvdp\" (UniqueName: \"kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp\") pod \"71a25257-e61a-493c-ae6b-42a29d97a6cd\" (UID: \"71a25257-e61a-493c-ae6b-42a29d97a6cd\") " Oct 01 08:58:12 crc kubenswrapper[4837]: I1001 08:58:12.610734 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp" (OuterVolumeSpecName: "kube-api-access-4pvdp") pod "71a25257-e61a-493c-ae6b-42a29d97a6cd" (UID: "71a25257-e61a-493c-ae6b-42a29d97a6cd"). InnerVolumeSpecName "kube-api-access-4pvdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:12 crc kubenswrapper[4837]: I1001 08:58:12.700132 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pvdp\" (UniqueName: \"kubernetes.io/projected/71a25257-e61a-493c-ae6b-42a29d97a6cd-kube-api-access-4pvdp\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:13 crc kubenswrapper[4837]: I1001 08:58:13.074060 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g5l26" event={"ID":"71a25257-e61a-493c-ae6b-42a29d97a6cd","Type":"ContainerDied","Data":"f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0"} Oct 01 08:58:13 crc kubenswrapper[4837]: I1001 08:58:13.074101 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6aad8a6a39a2154b60c19470175dd7335d5f10d48f385b848ec315c46c496f0" Oct 01 08:58:13 crc kubenswrapper[4837]: I1001 08:58:13.074124 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g5l26" Oct 01 08:58:17 crc kubenswrapper[4837]: I1001 08:58:17.135024 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.474560 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4fc3-account-create-4k5wb"] Oct 01 08:58:19 crc kubenswrapper[4837]: E1001 08:58:19.475567 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a25257-e61a-493c-ae6b-42a29d97a6cd" containerName="mariadb-database-create" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.475586 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a25257-e61a-493c-ae6b-42a29d97a6cd" containerName="mariadb-database-create" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.475786 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a25257-e61a-493c-ae6b-42a29d97a6cd" containerName="mariadb-database-create" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.476454 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.481526 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.487726 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fc3-account-create-4k5wb"] Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.529979 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlrcx\" (UniqueName: \"kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx\") pod \"keystone-4fc3-account-create-4k5wb\" (UID: \"8347ab54-35b3-41a2-a132-83654b5ad74c\") " pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.631810 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlrcx\" (UniqueName: \"kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx\") pod \"keystone-4fc3-account-create-4k5wb\" (UID: \"8347ab54-35b3-41a2-a132-83654b5ad74c\") " pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.653137 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlrcx\" (UniqueName: \"kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx\") pod \"keystone-4fc3-account-create-4k5wb\" (UID: \"8347ab54-35b3-41a2-a132-83654b5ad74c\") " pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:19 crc kubenswrapper[4837]: I1001 08:58:19.825788 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:20 crc kubenswrapper[4837]: I1001 08:58:20.302941 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fc3-account-create-4k5wb"] Oct 01 08:58:20 crc kubenswrapper[4837]: W1001 08:58:20.318591 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8347ab54_35b3_41a2_a132_83654b5ad74c.slice/crio-13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee WatchSource:0}: Error finding container 13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee: Status 404 returned error can't find the container with id 13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee Oct 01 08:58:21 crc kubenswrapper[4837]: I1001 08:58:21.160309 4837 generic.go:334] "Generic (PLEG): container finished" podID="8347ab54-35b3-41a2-a132-83654b5ad74c" containerID="ecb955d2f94ccc6c37f63bf596afd0d2e044080e0c1141afd28f384b7f7e2ed3" exitCode=0 Oct 01 08:58:21 crc kubenswrapper[4837]: I1001 08:58:21.160431 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc3-account-create-4k5wb" event={"ID":"8347ab54-35b3-41a2-a132-83654b5ad74c","Type":"ContainerDied","Data":"ecb955d2f94ccc6c37f63bf596afd0d2e044080e0c1141afd28f384b7f7e2ed3"} Oct 01 08:58:21 crc kubenswrapper[4837]: I1001 08:58:21.160746 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc3-account-create-4k5wb" event={"ID":"8347ab54-35b3-41a2-a132-83654b5ad74c","Type":"ContainerStarted","Data":"13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee"} Oct 01 08:58:22 crc kubenswrapper[4837]: I1001 08:58:22.613000 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:22 crc kubenswrapper[4837]: I1001 08:58:22.701196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlrcx\" (UniqueName: \"kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx\") pod \"8347ab54-35b3-41a2-a132-83654b5ad74c\" (UID: \"8347ab54-35b3-41a2-a132-83654b5ad74c\") " Oct 01 08:58:22 crc kubenswrapper[4837]: I1001 08:58:22.708968 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx" (OuterVolumeSpecName: "kube-api-access-nlrcx") pod "8347ab54-35b3-41a2-a132-83654b5ad74c" (UID: "8347ab54-35b3-41a2-a132-83654b5ad74c"). InnerVolumeSpecName "kube-api-access-nlrcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:22 crc kubenswrapper[4837]: I1001 08:58:22.803772 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlrcx\" (UniqueName: \"kubernetes.io/projected/8347ab54-35b3-41a2-a132-83654b5ad74c-kube-api-access-nlrcx\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:23 crc kubenswrapper[4837]: I1001 08:58:23.183627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fc3-account-create-4k5wb" event={"ID":"8347ab54-35b3-41a2-a132-83654b5ad74c","Type":"ContainerDied","Data":"13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee"} Oct 01 08:58:23 crc kubenswrapper[4837]: I1001 08:58:23.183668 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13e48b7c6cd3d8b84dea6c238c83834bdb2e68eb3096c935dabe9051d6bdcbee" Oct 01 08:58:23 crc kubenswrapper[4837]: I1001 08:58:23.183682 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fc3-account-create-4k5wb" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.868856 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dfl72"] Oct 01 08:58:24 crc kubenswrapper[4837]: E1001 08:58:24.869535 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8347ab54-35b3-41a2-a132-83654b5ad74c" containerName="mariadb-account-create" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.869551 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8347ab54-35b3-41a2-a132-83654b5ad74c" containerName="mariadb-account-create" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.869726 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8347ab54-35b3-41a2-a132-83654b5ad74c" containerName="mariadb-account-create" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.870449 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.876186 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.877484 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.877484 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qzdkz" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.880440 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.882182 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dfl72"] Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.946628 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.946813 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:24 crc kubenswrapper[4837]: I1001 08:58:24.946845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x8k8\" (UniqueName: \"kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.049043 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.049205 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.049251 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x8k8\" (UniqueName: \"kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.055560 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.063849 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.068374 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x8k8\" (UniqueName: \"kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8\") pod \"keystone-db-sync-dfl72\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.189049 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:25 crc kubenswrapper[4837]: I1001 08:58:25.692216 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dfl72"] Oct 01 08:58:25 crc kubenswrapper[4837]: W1001 08:58:25.700301 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0d2fd75_023a_4c79_beb1_e1a9393ba32e.slice/crio-88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943 WatchSource:0}: Error finding container 88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943: Status 404 returned error can't find the container with id 88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943 Oct 01 08:58:26 crc kubenswrapper[4837]: I1001 08:58:26.214981 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dfl72" event={"ID":"b0d2fd75-023a-4c79-beb1-e1a9393ba32e","Type":"ContainerStarted","Data":"88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943"} Oct 01 08:58:31 crc kubenswrapper[4837]: I1001 08:58:31.263974 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dfl72" event={"ID":"b0d2fd75-023a-4c79-beb1-e1a9393ba32e","Type":"ContainerStarted","Data":"1a7a4110f9731589a3b66021a05a5c4fa0b3cbc7e42312437b181b3af65f691c"} Oct 01 08:58:31 crc kubenswrapper[4837]: I1001 08:58:31.283902 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dfl72" podStartSLOduration=2.211618203 podStartE2EDuration="7.283877263s" podCreationTimestamp="2025-10-01 08:58:24 +0000 UTC" firstStartedPulling="2025-10-01 08:58:25.703591894 +0000 UTC m=+6762.545199389" lastFinishedPulling="2025-10-01 08:58:30.775850994 +0000 UTC m=+6767.617458449" observedRunningTime="2025-10-01 08:58:31.280345066 +0000 UTC m=+6768.121952531" watchObservedRunningTime="2025-10-01 08:58:31.283877263 +0000 UTC m=+6768.125484718" Oct 01 08:58:33 crc kubenswrapper[4837]: I1001 08:58:33.289819 4837 generic.go:334] "Generic (PLEG): container finished" podID="b0d2fd75-023a-4c79-beb1-e1a9393ba32e" containerID="1a7a4110f9731589a3b66021a05a5c4fa0b3cbc7e42312437b181b3af65f691c" exitCode=0 Oct 01 08:58:33 crc kubenswrapper[4837]: I1001 08:58:33.290003 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dfl72" event={"ID":"b0d2fd75-023a-4c79-beb1-e1a9393ba32e","Type":"ContainerDied","Data":"1a7a4110f9731589a3b66021a05a5c4fa0b3cbc7e42312437b181b3af65f691c"} Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.694468 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.736585 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x8k8\" (UniqueName: \"kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8\") pod \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.736744 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data\") pod \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.736864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle\") pod \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\" (UID: \"b0d2fd75-023a-4c79-beb1-e1a9393ba32e\") " Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.744447 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8" (OuterVolumeSpecName: "kube-api-access-2x8k8") pod "b0d2fd75-023a-4c79-beb1-e1a9393ba32e" (UID: "b0d2fd75-023a-4c79-beb1-e1a9393ba32e"). InnerVolumeSpecName "kube-api-access-2x8k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.774520 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0d2fd75-023a-4c79-beb1-e1a9393ba32e" (UID: "b0d2fd75-023a-4c79-beb1-e1a9393ba32e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.824819 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data" (OuterVolumeSpecName: "config-data") pod "b0d2fd75-023a-4c79-beb1-e1a9393ba32e" (UID: "b0d2fd75-023a-4c79-beb1-e1a9393ba32e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.839942 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x8k8\" (UniqueName: \"kubernetes.io/projected/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-kube-api-access-2x8k8\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.839999 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:34 crc kubenswrapper[4837]: I1001 08:58:34.840019 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0d2fd75-023a-4c79-beb1-e1a9393ba32e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.312199 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dfl72" event={"ID":"b0d2fd75-023a-4c79-beb1-e1a9393ba32e","Type":"ContainerDied","Data":"88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943"} Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.312509 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88e0ee69f6ecf3a05fa272bd960c2cef328c6a67f297801156e600df5342e943" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.312619 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dfl72" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.581437 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 08:58:35 crc kubenswrapper[4837]: E1001 08:58:35.581898 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0d2fd75-023a-4c79-beb1-e1a9393ba32e" containerName="keystone-db-sync" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.581921 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0d2fd75-023a-4c79-beb1-e1a9393ba32e" containerName="keystone-db-sync" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.582134 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0d2fd75-023a-4c79-beb1-e1a9393ba32e" containerName="keystone-db-sync" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.586035 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.610604 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.630144 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xf9vv"] Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.631358 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.634119 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.634252 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.635221 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.645207 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qzdkz" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.661356 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.661538 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.661641 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.661670 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.661762 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.687406 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xf9vv"] Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764073 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764129 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764191 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764260 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764359 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5gpv\" (UniqueName: \"kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764386 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764422 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764449 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764474 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764520 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.764546 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.765661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.766051 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.766381 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.766984 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.786672 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k\") pod \"dnsmasq-dns-f9c787459-zc4c9\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866527 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866642 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5gpv\" (UniqueName: \"kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866734 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.866763 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.870179 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.870236 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.870825 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.870985 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.871469 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.885453 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5gpv\" (UniqueName: \"kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv\") pod \"keystone-bootstrap-xf9vv\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.924875 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:35 crc kubenswrapper[4837]: I1001 08:58:35.972632 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:36 crc kubenswrapper[4837]: I1001 08:58:36.429221 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 08:58:36 crc kubenswrapper[4837]: W1001 08:58:36.430946 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4538c208_8c4d_4565_aa65_53cf2b137691.slice/crio-c994fef84f2912789316a0c9752aa28826a3af4590153f550a01c8a74272924c WatchSource:0}: Error finding container c994fef84f2912789316a0c9752aa28826a3af4590153f550a01c8a74272924c: Status 404 returned error can't find the container with id c994fef84f2912789316a0c9752aa28826a3af4590153f550a01c8a74272924c Oct 01 08:58:36 crc kubenswrapper[4837]: I1001 08:58:36.507925 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xf9vv"] Oct 01 08:58:36 crc kubenswrapper[4837]: W1001 08:58:36.512446 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda00ea37b_428f_435c_91af_f5f602e5e77d.slice/crio-a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75 WatchSource:0}: Error finding container a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75: Status 404 returned error can't find the container with id a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75 Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.336183 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xf9vv" event={"ID":"a00ea37b-428f-435c-91af-f5f602e5e77d","Type":"ContainerStarted","Data":"66169f6d97fb753d6c66b2b50344b91e550064a07cd0436defb4dd4fdca8c210"} Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.336777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xf9vv" event={"ID":"a00ea37b-428f-435c-91af-f5f602e5e77d","Type":"ContainerStarted","Data":"a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75"} Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.338222 4837 generic.go:334] "Generic (PLEG): container finished" podID="4538c208-8c4d-4565-aa65-53cf2b137691" containerID="a6e667d6ec6f4c61201bcce7746615daf9311ef8391ec4fcaf710f1efd9bc58a" exitCode=0 Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.338308 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" event={"ID":"4538c208-8c4d-4565-aa65-53cf2b137691","Type":"ContainerDied","Data":"a6e667d6ec6f4c61201bcce7746615daf9311ef8391ec4fcaf710f1efd9bc58a"} Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.338358 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" event={"ID":"4538c208-8c4d-4565-aa65-53cf2b137691","Type":"ContainerStarted","Data":"c994fef84f2912789316a0c9752aa28826a3af4590153f550a01c8a74272924c"} Oct 01 08:58:37 crc kubenswrapper[4837]: I1001 08:58:37.406193 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xf9vv" podStartSLOduration=2.406168525 podStartE2EDuration="2.406168525s" podCreationTimestamp="2025-10-01 08:58:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:58:37.388958863 +0000 UTC m=+6774.230566318" watchObservedRunningTime="2025-10-01 08:58:37.406168525 +0000 UTC m=+6774.247776000" Oct 01 08:58:38 crc kubenswrapper[4837]: I1001 08:58:38.348396 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" event={"ID":"4538c208-8c4d-4565-aa65-53cf2b137691","Type":"ContainerStarted","Data":"6cdacb8788bab3e066e8a50ad5011bc7b7cdee4430d7adae479cc687b136d1dd"} Oct 01 08:58:38 crc kubenswrapper[4837]: I1001 08:58:38.348870 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:38 crc kubenswrapper[4837]: I1001 08:58:38.381995 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" podStartSLOduration=3.381972766 podStartE2EDuration="3.381972766s" podCreationTimestamp="2025-10-01 08:58:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:58:38.376175744 +0000 UTC m=+6775.217783209" watchObservedRunningTime="2025-10-01 08:58:38.381972766 +0000 UTC m=+6775.223580221" Oct 01 08:58:41 crc kubenswrapper[4837]: I1001 08:58:41.380740 4837 generic.go:334] "Generic (PLEG): container finished" podID="a00ea37b-428f-435c-91af-f5f602e5e77d" containerID="66169f6d97fb753d6c66b2b50344b91e550064a07cd0436defb4dd4fdca8c210" exitCode=0 Oct 01 08:58:41 crc kubenswrapper[4837]: I1001 08:58:41.380850 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xf9vv" event={"ID":"a00ea37b-428f-435c-91af-f5f602e5e77d","Type":"ContainerDied","Data":"66169f6d97fb753d6c66b2b50344b91e550064a07cd0436defb4dd4fdca8c210"} Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.814733 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.929613 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.929746 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.929774 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5gpv\" (UniqueName: \"kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.929842 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.930484 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.930536 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys\") pod \"a00ea37b-428f-435c-91af-f5f602e5e77d\" (UID: \"a00ea37b-428f-435c-91af-f5f602e5e77d\") " Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.935764 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts" (OuterVolumeSpecName: "scripts") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.935897 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv" (OuterVolumeSpecName: "kube-api-access-c5gpv") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "kube-api-access-c5gpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.936296 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.939606 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.957342 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data" (OuterVolumeSpecName: "config-data") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:42 crc kubenswrapper[4837]: I1001 08:58:42.959650 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a00ea37b-428f-435c-91af-f5f602e5e77d" (UID: "a00ea37b-428f-435c-91af-f5f602e5e77d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032789 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032821 4837 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032830 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032839 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032848 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5gpv\" (UniqueName: \"kubernetes.io/projected/a00ea37b-428f-435c-91af-f5f602e5e77d-kube-api-access-c5gpv\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.032857 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a00ea37b-428f-435c-91af-f5f602e5e77d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.406885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xf9vv" event={"ID":"a00ea37b-428f-435c-91af-f5f602e5e77d","Type":"ContainerDied","Data":"a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75"} Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.406946 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a93ccba109e8cf37c9423912394677dc10514a8ca26d89a08186db5189e27d75" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.406972 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xf9vv" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.506968 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xf9vv"] Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.514539 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xf9vv"] Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.614612 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wnz4p"] Oct 01 08:58:43 crc kubenswrapper[4837]: E1001 08:58:43.615020 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00ea37b-428f-435c-91af-f5f602e5e77d" containerName="keystone-bootstrap" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.615039 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00ea37b-428f-435c-91af-f5f602e5e77d" containerName="keystone-bootstrap" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.615217 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00ea37b-428f-435c-91af-f5f602e5e77d" containerName="keystone-bootstrap" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.615817 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.618527 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.619015 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.622030 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qzdkz" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.623079 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.634175 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wnz4p"] Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746173 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746335 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746518 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtqj\" (UniqueName: \"kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746725 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.746787 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.831258 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00ea37b-428f-435c-91af-f5f602e5e77d" path="/var/lib/kubelet/pods/a00ea37b-428f-435c-91af-f5f602e5e77d/volumes" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.848899 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.848972 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.849032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.849061 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.849088 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.849119 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtqj\" (UniqueName: \"kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.852056 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.852086 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.852241 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.865077 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.865320 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.865608 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.866280 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.872921 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.879900 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtqj\" (UniqueName: \"kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj\") pod \"keystone-bootstrap-wnz4p\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.941161 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qzdkz" Oct 01 08:58:43 crc kubenswrapper[4837]: I1001 08:58:43.950988 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:44 crc kubenswrapper[4837]: I1001 08:58:44.435550 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wnz4p"] Oct 01 08:58:45 crc kubenswrapper[4837]: I1001 08:58:45.446355 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnz4p" event={"ID":"186b2215-73a0-4142-b1d3-139f6b72a2b1","Type":"ContainerStarted","Data":"c3f4acfe0747d97132b8302af5952f78b17b79af690b8c297660fb212f1dad2a"} Oct 01 08:58:45 crc kubenswrapper[4837]: I1001 08:58:45.446965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnz4p" event={"ID":"186b2215-73a0-4142-b1d3-139f6b72a2b1","Type":"ContainerStarted","Data":"35731f989e03c98aa38489324d0106cb72740340c541b8bcbc79eb6531c1cc9f"} Oct 01 08:58:45 crc kubenswrapper[4837]: I1001 08:58:45.471374 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wnz4p" podStartSLOduration=2.471357265 podStartE2EDuration="2.471357265s" podCreationTimestamp="2025-10-01 08:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:58:45.467426139 +0000 UTC m=+6782.309033594" watchObservedRunningTime="2025-10-01 08:58:45.471357265 +0000 UTC m=+6782.312964720" Oct 01 08:58:45 crc kubenswrapper[4837]: I1001 08:58:45.927009 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.020993 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.021235 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="dnsmasq-dns" containerID="cri-o://2e31be9975305e39a5602ea5a2b1bfee78f1113567b1ccb6803ce3073163ed37" gracePeriod=10 Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.464913 4837 generic.go:334] "Generic (PLEG): container finished" podID="23a56f34-a1fd-42c8-a634-bae396251457" containerID="2e31be9975305e39a5602ea5a2b1bfee78f1113567b1ccb6803ce3073163ed37" exitCode=0 Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.464972 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" event={"ID":"23a56f34-a1fd-42c8-a634-bae396251457","Type":"ContainerDied","Data":"2e31be9975305e39a5602ea5a2b1bfee78f1113567b1ccb6803ce3073163ed37"} Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.465016 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" event={"ID":"23a56f34-a1fd-42c8-a634-bae396251457","Type":"ContainerDied","Data":"ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec"} Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.465030 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccbfb66a33b84e475bf7b131a395a5359d67168b2969547558493598fe24f6ec" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.507592 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.609575 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.609656 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxzvg\" (UniqueName: \"kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.609710 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.609783 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.609880 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.629043 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg" (OuterVolumeSpecName: "kube-api-access-qxzvg") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457"). InnerVolumeSpecName "kube-api-access-qxzvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.658522 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.662709 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:58:46 crc kubenswrapper[4837]: E1001 08:58:46.669883 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config podName:23a56f34-a1fd-42c8-a634-bae396251457 nodeName:}" failed. No retries permitted until 2025-10-01 08:58:47.169850536 +0000 UTC m=+6784.011457991 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457") : error deleting /var/lib/kubelet/pods/23a56f34-a1fd-42c8-a634-bae396251457/volume-subpaths: remove /var/lib/kubelet/pods/23a56f34-a1fd-42c8-a634-bae396251457/volume-subpaths: no such file or directory Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.672107 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.711950 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.712007 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.712020 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:46 crc kubenswrapper[4837]: I1001 08:58:46.712035 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxzvg\" (UniqueName: \"kubernetes.io/projected/23a56f34-a1fd-42c8-a634-bae396251457-kube-api-access-qxzvg\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.223593 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") pod \"23a56f34-a1fd-42c8-a634-bae396251457\" (UID: \"23a56f34-a1fd-42c8-a634-bae396251457\") " Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.224364 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config" (OuterVolumeSpecName: "config") pod "23a56f34-a1fd-42c8-a634-bae396251457" (UID: "23a56f34-a1fd-42c8-a634-bae396251457"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.225023 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23a56f34-a1fd-42c8-a634-bae396251457-config\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.475509 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744d59f9-95kt7" Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.514468 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.522917 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7744d59f9-95kt7"] Oct 01 08:58:47 crc kubenswrapper[4837]: I1001 08:58:47.834542 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23a56f34-a1fd-42c8-a634-bae396251457" path="/var/lib/kubelet/pods/23a56f34-a1fd-42c8-a634-bae396251457/volumes" Oct 01 08:58:48 crc kubenswrapper[4837]: I1001 08:58:48.487171 4837 generic.go:334] "Generic (PLEG): container finished" podID="186b2215-73a0-4142-b1d3-139f6b72a2b1" containerID="c3f4acfe0747d97132b8302af5952f78b17b79af690b8c297660fb212f1dad2a" exitCode=0 Oct 01 08:58:48 crc kubenswrapper[4837]: I1001 08:58:48.487244 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnz4p" event={"ID":"186b2215-73a0-4142-b1d3-139f6b72a2b1","Type":"ContainerDied","Data":"c3f4acfe0747d97132b8302af5952f78b17b79af690b8c297660fb212f1dad2a"} Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.892427 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.981992 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.982054 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.982098 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.982232 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.982263 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtqj\" (UniqueName: \"kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.982280 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data\") pod \"186b2215-73a0-4142-b1d3-139f6b72a2b1\" (UID: \"186b2215-73a0-4142-b1d3-139f6b72a2b1\") " Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.988025 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj" (OuterVolumeSpecName: "kube-api-access-fwtqj") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "kube-api-access-fwtqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.988101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.988509 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts" (OuterVolumeSpecName: "scripts") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:49 crc kubenswrapper[4837]: I1001 08:58:49.993922 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.009225 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data" (OuterVolumeSpecName: "config-data") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.012592 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "186b2215-73a0-4142-b1d3-139f6b72a2b1" (UID: "186b2215-73a0-4142-b1d3-139f6b72a2b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084753 4837 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084799 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtqj\" (UniqueName: \"kubernetes.io/projected/186b2215-73a0-4142-b1d3-139f6b72a2b1-kube-api-access-fwtqj\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084814 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084824 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084835 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.084844 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/186b2215-73a0-4142-b1d3-139f6b72a2b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.507149 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wnz4p" event={"ID":"186b2215-73a0-4142-b1d3-139f6b72a2b1","Type":"ContainerDied","Data":"35731f989e03c98aa38489324d0106cb72740340c541b8bcbc79eb6531c1cc9f"} Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.507793 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35731f989e03c98aa38489324d0106cb72740340c541b8bcbc79eb6531c1cc9f" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.507248 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wnz4p" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.618908 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9788869bb-qfvkm"] Oct 01 08:58:50 crc kubenswrapper[4837]: E1001 08:58:50.619294 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="init" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.619311 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="init" Oct 01 08:58:50 crc kubenswrapper[4837]: E1001 08:58:50.619330 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="186b2215-73a0-4142-b1d3-139f6b72a2b1" containerName="keystone-bootstrap" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.619336 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="186b2215-73a0-4142-b1d3-139f6b72a2b1" containerName="keystone-bootstrap" Oct 01 08:58:50 crc kubenswrapper[4837]: E1001 08:58:50.619345 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="dnsmasq-dns" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.619351 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="dnsmasq-dns" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.619516 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a56f34-a1fd-42c8-a634-bae396251457" containerName="dnsmasq-dns" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.619528 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="186b2215-73a0-4142-b1d3-139f6b72a2b1" containerName="keystone-bootstrap" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.620157 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.623473 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.623639 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.623836 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qzdkz" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.624165 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.624282 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.624453 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.639999 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9788869bb-qfvkm"] Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.697844 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-config-data\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.697898 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-fernet-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.697929 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-credential-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.698167 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-combined-ca-bundle\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.698256 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-public-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.698284 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dxb\" (UniqueName: \"kubernetes.io/projected/d4d8d5b0-1282-4dff-92ff-5a79de21571c-kube-api-access-96dxb\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.698380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-internal-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.698456 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-scripts\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800342 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-combined-ca-bundle\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800402 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-public-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dxb\" (UniqueName: \"kubernetes.io/projected/d4d8d5b0-1282-4dff-92ff-5a79de21571c-kube-api-access-96dxb\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800452 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-internal-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800477 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-scripts\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-config-data\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-fernet-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.800573 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-credential-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.804355 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-credential-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.804502 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-combined-ca-bundle\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.806957 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-scripts\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.806957 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-public-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.809199 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-config-data\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.809894 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-fernet-keys\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.810153 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4d8d5b0-1282-4dff-92ff-5a79de21571c-internal-tls-certs\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.821648 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dxb\" (UniqueName: \"kubernetes.io/projected/d4d8d5b0-1282-4dff-92ff-5a79de21571c-kube-api-access-96dxb\") pod \"keystone-9788869bb-qfvkm\" (UID: \"d4d8d5b0-1282-4dff-92ff-5a79de21571c\") " pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:50 crc kubenswrapper[4837]: I1001 08:58:50.959971 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:51 crc kubenswrapper[4837]: I1001 08:58:51.423870 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9788869bb-qfvkm"] Oct 01 08:58:51 crc kubenswrapper[4837]: W1001 08:58:51.427782 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4d8d5b0_1282_4dff_92ff_5a79de21571c.slice/crio-330273df2b02a421d8b74743ea91332fe3a43d7f0a1555ac0052c9735df460b2 WatchSource:0}: Error finding container 330273df2b02a421d8b74743ea91332fe3a43d7f0a1555ac0052c9735df460b2: Status 404 returned error can't find the container with id 330273df2b02a421d8b74743ea91332fe3a43d7f0a1555ac0052c9735df460b2 Oct 01 08:58:51 crc kubenswrapper[4837]: I1001 08:58:51.517461 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9788869bb-qfvkm" event={"ID":"d4d8d5b0-1282-4dff-92ff-5a79de21571c","Type":"ContainerStarted","Data":"330273df2b02a421d8b74743ea91332fe3a43d7f0a1555ac0052c9735df460b2"} Oct 01 08:58:52 crc kubenswrapper[4837]: I1001 08:58:52.529813 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9788869bb-qfvkm" event={"ID":"d4d8d5b0-1282-4dff-92ff-5a79de21571c","Type":"ContainerStarted","Data":"46766fbbfd9cebbb1b3ad87ef90c3c661392eb1a17c3a4b091ff5ddd3fafa453"} Oct 01 08:58:52 crc kubenswrapper[4837]: I1001 08:58:52.530384 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:58:52 crc kubenswrapper[4837]: I1001 08:58:52.567871 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9788869bb-qfvkm" podStartSLOduration=2.5678419999999997 podStartE2EDuration="2.567842s" podCreationTimestamp="2025-10-01 08:58:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 08:58:52.558682784 +0000 UTC m=+6789.400290249" watchObservedRunningTime="2025-10-01 08:58:52.567842 +0000 UTC m=+6789.409449485" Oct 01 08:58:53 crc kubenswrapper[4837]: I1001 08:58:53.079769 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:58:53 crc kubenswrapper[4837]: I1001 08:58:53.079861 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:59:22 crc kubenswrapper[4837]: I1001 08:59:22.506666 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-9788869bb-qfvkm" Oct 01 08:59:23 crc kubenswrapper[4837]: I1001 08:59:23.079560 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:59:23 crc kubenswrapper[4837]: I1001 08:59:23.079635 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.577199 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.579271 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.582164 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.582577 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-82jw6" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.582848 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.588560 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.623846 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7j7\" (UniqueName: \"kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.623964 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.624017 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.624098 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.726500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.726732 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.726826 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7j7\" (UniqueName: \"kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.726917 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.728170 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.736530 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.743629 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.747548 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7j7\" (UniqueName: \"kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7\") pod \"openstackclient\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " pod="openstack/openstackclient" Oct 01 08:59:24 crc kubenswrapper[4837]: I1001 08:59:24.929955 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 08:59:25 crc kubenswrapper[4837]: I1001 08:59:25.412381 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 08:59:25 crc kubenswrapper[4837]: I1001 08:59:25.888580 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b327f5dc-78e7-412c-a281-d281511fefa5","Type":"ContainerStarted","Data":"0cce82aa885899e50ddfdadf0932f615e7ba2cdfb9dab1fe874209b7389be821"} Oct 01 08:59:39 crc kubenswrapper[4837]: I1001 08:59:39.011948 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b327f5dc-78e7-412c-a281-d281511fefa5","Type":"ContainerStarted","Data":"b7a9137cf9d14a8e88f10c96a379d917d6fc6c0b280bd6a9405e8d1a81622fe2"} Oct 01 08:59:39 crc kubenswrapper[4837]: I1001 08:59:39.037735 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.354324318 podStartE2EDuration="15.037675773s" podCreationTimestamp="2025-10-01 08:59:24 +0000 UTC" firstStartedPulling="2025-10-01 08:59:25.432219427 +0000 UTC m=+6822.273826882" lastFinishedPulling="2025-10-01 08:59:38.115570872 +0000 UTC m=+6834.957178337" observedRunningTime="2025-10-01 08:59:39.034366252 +0000 UTC m=+6835.875973747" watchObservedRunningTime="2025-10-01 08:59:39.037675773 +0000 UTC m=+6835.879283268" Oct 01 08:59:53 crc kubenswrapper[4837]: I1001 08:59:53.080124 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 08:59:53 crc kubenswrapper[4837]: I1001 08:59:53.081017 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 08:59:53 crc kubenswrapper[4837]: I1001 08:59:53.081106 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 08:59:53 crc kubenswrapper[4837]: I1001 08:59:53.082316 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 08:59:53 crc kubenswrapper[4837]: I1001 08:59:53.082444 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5" gracePeriod=600 Oct 01 08:59:54 crc kubenswrapper[4837]: I1001 08:59:54.186083 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5" exitCode=0 Oct 01 08:59:54 crc kubenswrapper[4837]: I1001 08:59:54.186177 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5"} Oct 01 08:59:54 crc kubenswrapper[4837]: I1001 08:59:54.187094 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6"} Oct 01 08:59:54 crc kubenswrapper[4837]: I1001 08:59:54.187125 4837 scope.go:117] "RemoveContainer" containerID="8b572f0c8b1d4c1c93ebc21cdf4d0a0f7fe815020240504f26c24cd80ca42921" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.193732 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn"] Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.195780 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.199684 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.200052 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.212542 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn"] Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.354394 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.354598 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdfnz\" (UniqueName: \"kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.354720 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.457377 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.457985 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdfnz\" (UniqueName: \"kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.458169 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.458620 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.465539 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.487073 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdfnz\" (UniqueName: \"kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz\") pod \"collect-profiles-29321820-rg2tn\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:00 crc kubenswrapper[4837]: I1001 09:00:00.523291 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:01 crc kubenswrapper[4837]: I1001 09:00:01.012447 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn"] Oct 01 09:00:01 crc kubenswrapper[4837]: I1001 09:00:01.264703 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" event={"ID":"2d2727e4-0222-4d79-b696-eff0350a53eb","Type":"ContainerStarted","Data":"2136ad3faee1291bcf38372fc195fe002b174a7830e0534ec0f67dfa651a1954"} Oct 01 09:00:01 crc kubenswrapper[4837]: I1001 09:00:01.264751 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" event={"ID":"2d2727e4-0222-4d79-b696-eff0350a53eb","Type":"ContainerStarted","Data":"00d0de1ca356096b130c15ef5002f25cef2bfc9c0055fdec3a1b7c0fae957f42"} Oct 01 09:00:01 crc kubenswrapper[4837]: I1001 09:00:01.300466 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" podStartSLOduration=1.300443953 podStartE2EDuration="1.300443953s" podCreationTimestamp="2025-10-01 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:00:01.28406445 +0000 UTC m=+6858.125671905" watchObservedRunningTime="2025-10-01 09:00:01.300443953 +0000 UTC m=+6858.142051408" Oct 01 09:00:02 crc kubenswrapper[4837]: I1001 09:00:02.280020 4837 generic.go:334] "Generic (PLEG): container finished" podID="2d2727e4-0222-4d79-b696-eff0350a53eb" containerID="2136ad3faee1291bcf38372fc195fe002b174a7830e0534ec0f67dfa651a1954" exitCode=0 Oct 01 09:00:02 crc kubenswrapper[4837]: I1001 09:00:02.280077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" event={"ID":"2d2727e4-0222-4d79-b696-eff0350a53eb","Type":"ContainerDied","Data":"2136ad3faee1291bcf38372fc195fe002b174a7830e0534ec0f67dfa651a1954"} Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.629884 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.724929 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdfnz\" (UniqueName: \"kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz\") pod \"2d2727e4-0222-4d79-b696-eff0350a53eb\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.724996 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume\") pod \"2d2727e4-0222-4d79-b696-eff0350a53eb\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.725035 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume\") pod \"2d2727e4-0222-4d79-b696-eff0350a53eb\" (UID: \"2d2727e4-0222-4d79-b696-eff0350a53eb\") " Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.726093 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "2d2727e4-0222-4d79-b696-eff0350a53eb" (UID: "2d2727e4-0222-4d79-b696-eff0350a53eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.731357 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2d2727e4-0222-4d79-b696-eff0350a53eb" (UID: "2d2727e4-0222-4d79-b696-eff0350a53eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.732992 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz" (OuterVolumeSpecName: "kube-api-access-xdfnz") pod "2d2727e4-0222-4d79-b696-eff0350a53eb" (UID: "2d2727e4-0222-4d79-b696-eff0350a53eb"). InnerVolumeSpecName "kube-api-access-xdfnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.827764 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdfnz\" (UniqueName: \"kubernetes.io/projected/2d2727e4-0222-4d79-b696-eff0350a53eb-kube-api-access-xdfnz\") on node \"crc\" DevicePath \"\"" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.827796 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d2727e4-0222-4d79-b696-eff0350a53eb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:00:03 crc kubenswrapper[4837]: I1001 09:00:03.827808 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d2727e4-0222-4d79-b696-eff0350a53eb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:00:04 crc kubenswrapper[4837]: I1001 09:00:04.300921 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" event={"ID":"2d2727e4-0222-4d79-b696-eff0350a53eb","Type":"ContainerDied","Data":"00d0de1ca356096b130c15ef5002f25cef2bfc9c0055fdec3a1b7c0fae957f42"} Oct 01 09:00:04 crc kubenswrapper[4837]: I1001 09:00:04.300986 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00d0de1ca356096b130c15ef5002f25cef2bfc9c0055fdec3a1b7c0fae957f42" Oct 01 09:00:04 crc kubenswrapper[4837]: I1001 09:00:04.301021 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn" Oct 01 09:00:04 crc kubenswrapper[4837]: I1001 09:00:04.367858 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c"] Oct 01 09:00:04 crc kubenswrapper[4837]: I1001 09:00:04.375153 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321775-6z86c"] Oct 01 09:00:05 crc kubenswrapper[4837]: I1001 09:00:05.827964 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="261058d3-6d4a-4612-a6dc-6b0882515208" path="/var/lib/kubelet/pods/261058d3-6d4a-4612-a6dc-6b0882515208/volumes" Oct 01 09:00:06 crc kubenswrapper[4837]: I1001 09:00:06.519458 4837 scope.go:117] "RemoveContainer" containerID="775abe43fd64391df42b55ef1cabcdc7f23f6e10ef7f845ed4d53d3b4a37f624" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.159278 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29321821-xqq67"] Oct 01 09:01:00 crc kubenswrapper[4837]: E1001 09:01:00.160536 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2727e4-0222-4d79-b696-eff0350a53eb" containerName="collect-profiles" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.160557 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2727e4-0222-4d79-b696-eff0350a53eb" containerName="collect-profiles" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.160789 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d2727e4-0222-4d79-b696-eff0350a53eb" containerName="collect-profiles" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.161574 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.169187 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321821-xqq67"] Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.253864 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.253937 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.254031 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6lz7\" (UniqueName: \"kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.254128 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.356492 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6lz7\" (UniqueName: \"kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.356575 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.356625 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.356652 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.368331 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.368450 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.369140 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.381046 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6lz7\" (UniqueName: \"kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7\") pod \"keystone-cron-29321821-xqq67\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.487485 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:00 crc kubenswrapper[4837]: I1001 09:01:00.976034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321821-xqq67"] Oct 01 09:01:00 crc kubenswrapper[4837]: W1001 09:01:00.982785 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc041921_e6b3_4855_9685_8855f4efe2db.slice/crio-bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20 WatchSource:0}: Error finding container bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20: Status 404 returned error can't find the container with id bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20 Oct 01 09:01:01 crc kubenswrapper[4837]: I1001 09:01:01.915805 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321821-xqq67" event={"ID":"bc041921-e6b3-4855-9685-8855f4efe2db","Type":"ContainerStarted","Data":"ce791cdd4274f0010c44ec2e426450a56137c135003e2e5a1437625ea6e1dc1d"} Oct 01 09:01:01 crc kubenswrapper[4837]: I1001 09:01:01.916374 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321821-xqq67" event={"ID":"bc041921-e6b3-4855-9685-8855f4efe2db","Type":"ContainerStarted","Data":"bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20"} Oct 01 09:01:01 crc kubenswrapper[4837]: I1001 09:01:01.939030 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29321821-xqq67" podStartSLOduration=1.939009207 podStartE2EDuration="1.939009207s" podCreationTimestamp="2025-10-01 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:01:01.931773648 +0000 UTC m=+6918.773381103" watchObservedRunningTime="2025-10-01 09:01:01.939009207 +0000 UTC m=+6918.780616662" Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.486525 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-42cq8"] Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.487819 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.511553 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-42cq8"] Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.601052 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm9lf\" (UniqueName: \"kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf\") pod \"barbican-db-create-42cq8\" (UID: \"7892968b-bed5-4332-b510-12f2b12caefd\") " pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.703181 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm9lf\" (UniqueName: \"kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf\") pod \"barbican-db-create-42cq8\" (UID: \"7892968b-bed5-4332-b510-12f2b12caefd\") " pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.735666 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm9lf\" (UniqueName: \"kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf\") pod \"barbican-db-create-42cq8\" (UID: \"7892968b-bed5-4332-b510-12f2b12caefd\") " pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:02 crc kubenswrapper[4837]: I1001 09:01:02.812158 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.416523 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-42cq8"] Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.955353 4837 generic.go:334] "Generic (PLEG): container finished" podID="bc041921-e6b3-4855-9685-8855f4efe2db" containerID="ce791cdd4274f0010c44ec2e426450a56137c135003e2e5a1437625ea6e1dc1d" exitCode=0 Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.955411 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321821-xqq67" event={"ID":"bc041921-e6b3-4855-9685-8855f4efe2db","Type":"ContainerDied","Data":"ce791cdd4274f0010c44ec2e426450a56137c135003e2e5a1437625ea6e1dc1d"} Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.957557 4837 generic.go:334] "Generic (PLEG): container finished" podID="7892968b-bed5-4332-b510-12f2b12caefd" containerID="636dde0f86ee0b9dfdaa5155ee41599f956c39e647ffe136f2a8680a46555cea" exitCode=0 Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.957604 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-42cq8" event={"ID":"7892968b-bed5-4332-b510-12f2b12caefd","Type":"ContainerDied","Data":"636dde0f86ee0b9dfdaa5155ee41599f956c39e647ffe136f2a8680a46555cea"} Oct 01 09:01:03 crc kubenswrapper[4837]: I1001 09:01:03.957631 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-42cq8" event={"ID":"7892968b-bed5-4332-b510-12f2b12caefd","Type":"ContainerStarted","Data":"5716f6ac88ddadba5eabce3ccfd78b7d036663929154e38ed0f56d725b9327a1"} Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.408907 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.413966 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.557907 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys\") pod \"bc041921-e6b3-4855-9685-8855f4efe2db\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.558031 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data\") pod \"bc041921-e6b3-4855-9685-8855f4efe2db\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.558264 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm9lf\" (UniqueName: \"kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf\") pod \"7892968b-bed5-4332-b510-12f2b12caefd\" (UID: \"7892968b-bed5-4332-b510-12f2b12caefd\") " Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.558342 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle\") pod \"bc041921-e6b3-4855-9685-8855f4efe2db\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.558472 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6lz7\" (UniqueName: \"kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7\") pod \"bc041921-e6b3-4855-9685-8855f4efe2db\" (UID: \"bc041921-e6b3-4855-9685-8855f4efe2db\") " Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.563638 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bc041921-e6b3-4855-9685-8855f4efe2db" (UID: "bc041921-e6b3-4855-9685-8855f4efe2db"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.571422 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf" (OuterVolumeSpecName: "kube-api-access-qm9lf") pod "7892968b-bed5-4332-b510-12f2b12caefd" (UID: "7892968b-bed5-4332-b510-12f2b12caefd"). InnerVolumeSpecName "kube-api-access-qm9lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.571864 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7" (OuterVolumeSpecName: "kube-api-access-s6lz7") pod "bc041921-e6b3-4855-9685-8855f4efe2db" (UID: "bc041921-e6b3-4855-9685-8855f4efe2db"). InnerVolumeSpecName "kube-api-access-s6lz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.616480 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc041921-e6b3-4855-9685-8855f4efe2db" (UID: "bc041921-e6b3-4855-9685-8855f4efe2db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.633918 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data" (OuterVolumeSpecName: "config-data") pod "bc041921-e6b3-4855-9685-8855f4efe2db" (UID: "bc041921-e6b3-4855-9685-8855f4efe2db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.660149 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6lz7\" (UniqueName: \"kubernetes.io/projected/bc041921-e6b3-4855-9685-8855f4efe2db-kube-api-access-s6lz7\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.660185 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.660198 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.660210 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm9lf\" (UniqueName: \"kubernetes.io/projected/7892968b-bed5-4332-b510-12f2b12caefd-kube-api-access-qm9lf\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.660279 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc041921-e6b3-4855-9685-8855f4efe2db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.982111 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-42cq8" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.982106 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-42cq8" event={"ID":"7892968b-bed5-4332-b510-12f2b12caefd","Type":"ContainerDied","Data":"5716f6ac88ddadba5eabce3ccfd78b7d036663929154e38ed0f56d725b9327a1"} Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.982161 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5716f6ac88ddadba5eabce3ccfd78b7d036663929154e38ed0f56d725b9327a1" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.984761 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321821-xqq67" event={"ID":"bc041921-e6b3-4855-9685-8855f4efe2db","Type":"ContainerDied","Data":"bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20"} Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.984782 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf9549b1f7ec62bd97bef1086d18d6eacb90e2905fcd0fd3b3ebf089b07fbb20" Oct 01 09:01:05 crc kubenswrapper[4837]: I1001 09:01:05.984830 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321821-xqq67" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.496238 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-633e-account-create-hdn2v"] Oct 01 09:01:12 crc kubenswrapper[4837]: E1001 09:01:12.497488 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7892968b-bed5-4332-b510-12f2b12caefd" containerName="mariadb-database-create" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.497515 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7892968b-bed5-4332-b510-12f2b12caefd" containerName="mariadb-database-create" Oct 01 09:01:12 crc kubenswrapper[4837]: E1001 09:01:12.497547 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc041921-e6b3-4855-9685-8855f4efe2db" containerName="keystone-cron" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.497561 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc041921-e6b3-4855-9685-8855f4efe2db" containerName="keystone-cron" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.497993 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7892968b-bed5-4332-b510-12f2b12caefd" containerName="mariadb-database-create" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.498020 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc041921-e6b3-4855-9685-8855f4efe2db" containerName="keystone-cron" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.499032 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.504164 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.512729 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-633e-account-create-hdn2v"] Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.597079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdg92\" (UniqueName: \"kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92\") pod \"barbican-633e-account-create-hdn2v\" (UID: \"0196e178-95f4-4e0f-b09b-f420c71859bc\") " pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.698335 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdg92\" (UniqueName: \"kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92\") pod \"barbican-633e-account-create-hdn2v\" (UID: \"0196e178-95f4-4e0f-b09b-f420c71859bc\") " pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.720596 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdg92\" (UniqueName: \"kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92\") pod \"barbican-633e-account-create-hdn2v\" (UID: \"0196e178-95f4-4e0f-b09b-f420c71859bc\") " pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:12 crc kubenswrapper[4837]: I1001 09:01:12.828562 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:13 crc kubenswrapper[4837]: I1001 09:01:13.326419 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-633e-account-create-hdn2v"] Oct 01 09:01:14 crc kubenswrapper[4837]: I1001 09:01:14.066210 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-633e-account-create-hdn2v" event={"ID":"0196e178-95f4-4e0f-b09b-f420c71859bc","Type":"ContainerStarted","Data":"4fe01c92544f4d065eaea42022911abdba66f825dc77d9c6e1ba7233be6ed99b"} Oct 01 09:01:14 crc kubenswrapper[4837]: I1001 09:01:14.066539 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-633e-account-create-hdn2v" event={"ID":"0196e178-95f4-4e0f-b09b-f420c71859bc","Type":"ContainerStarted","Data":"cecfcac1922c71f808e922a090023308ce3234e7aa58b69e5f37203385760855"} Oct 01 09:01:14 crc kubenswrapper[4837]: I1001 09:01:14.090522 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-633e-account-create-hdn2v" podStartSLOduration=2.090487544 podStartE2EDuration="2.090487544s" podCreationTimestamp="2025-10-01 09:01:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:01:14.085215376 +0000 UTC m=+6930.926822871" watchObservedRunningTime="2025-10-01 09:01:14.090487544 +0000 UTC m=+6930.932095039" Oct 01 09:01:15 crc kubenswrapper[4837]: I1001 09:01:15.077112 4837 generic.go:334] "Generic (PLEG): container finished" podID="0196e178-95f4-4e0f-b09b-f420c71859bc" containerID="4fe01c92544f4d065eaea42022911abdba66f825dc77d9c6e1ba7233be6ed99b" exitCode=0 Oct 01 09:01:15 crc kubenswrapper[4837]: I1001 09:01:15.077121 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-633e-account-create-hdn2v" event={"ID":"0196e178-95f4-4e0f-b09b-f420c71859bc","Type":"ContainerDied","Data":"4fe01c92544f4d065eaea42022911abdba66f825dc77d9c6e1ba7233be6ed99b"} Oct 01 09:01:16 crc kubenswrapper[4837]: I1001 09:01:16.477109 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:16 crc kubenswrapper[4837]: I1001 09:01:16.583965 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdg92\" (UniqueName: \"kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92\") pod \"0196e178-95f4-4e0f-b09b-f420c71859bc\" (UID: \"0196e178-95f4-4e0f-b09b-f420c71859bc\") " Oct 01 09:01:16 crc kubenswrapper[4837]: I1001 09:01:16.590837 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92" (OuterVolumeSpecName: "kube-api-access-kdg92") pod "0196e178-95f4-4e0f-b09b-f420c71859bc" (UID: "0196e178-95f4-4e0f-b09b-f420c71859bc"). InnerVolumeSpecName "kube-api-access-kdg92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:16 crc kubenswrapper[4837]: I1001 09:01:16.686861 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdg92\" (UniqueName: \"kubernetes.io/projected/0196e178-95f4-4e0f-b09b-f420c71859bc-kube-api-access-kdg92\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.101492 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-633e-account-create-hdn2v" event={"ID":"0196e178-95f4-4e0f-b09b-f420c71859bc","Type":"ContainerDied","Data":"cecfcac1922c71f808e922a090023308ce3234e7aa58b69e5f37203385760855"} Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.101550 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cecfcac1922c71f808e922a090023308ce3234e7aa58b69e5f37203385760855" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.101611 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-633e-account-create-hdn2v" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.802776 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-mbxg6"] Oct 01 09:01:17 crc kubenswrapper[4837]: E1001 09:01:17.803198 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0196e178-95f4-4e0f-b09b-f420c71859bc" containerName="mariadb-account-create" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.803214 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0196e178-95f4-4e0f-b09b-f420c71859bc" containerName="mariadb-account-create" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.803412 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0196e178-95f4-4e0f-b09b-f420c71859bc" containerName="mariadb-account-create" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.804109 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.809771 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dvvbm" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.809982 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.831229 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mbxg6"] Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.912242 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89c5k\" (UniqueName: \"kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.913215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:17 crc kubenswrapper[4837]: I1001 09:01:17.913832 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.015297 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89c5k\" (UniqueName: \"kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.015414 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.015463 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.019508 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.074566 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.094875 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89c5k\" (UniqueName: \"kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k\") pod \"barbican-db-sync-mbxg6\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.135415 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.643230 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mbxg6"] Oct 01 09:01:18 crc kubenswrapper[4837]: I1001 09:01:18.651644 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:01:19 crc kubenswrapper[4837]: I1001 09:01:19.123748 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbxg6" event={"ID":"9ef08c36-8f34-4174-bc6f-d6bdaa96a498","Type":"ContainerStarted","Data":"7e8cc5e0f38b5e1918bf1754ce02e6c6148f54d64983df8264594e89052e01fc"} Oct 01 09:01:24 crc kubenswrapper[4837]: I1001 09:01:24.168504 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbxg6" event={"ID":"9ef08c36-8f34-4174-bc6f-d6bdaa96a498","Type":"ContainerStarted","Data":"8f5e9afc1bf928299fb9aeacbeb044a6916edd586d2dee2febe5daba3bb7575b"} Oct 01 09:01:24 crc kubenswrapper[4837]: I1001 09:01:24.190619 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-mbxg6" podStartSLOduration=2.656826321 podStartE2EDuration="7.190597212s" podCreationTimestamp="2025-10-01 09:01:17 +0000 UTC" firstStartedPulling="2025-10-01 09:01:18.651382913 +0000 UTC m=+6935.492990368" lastFinishedPulling="2025-10-01 09:01:23.185153804 +0000 UTC m=+6940.026761259" observedRunningTime="2025-10-01 09:01:24.182311868 +0000 UTC m=+6941.023919363" watchObservedRunningTime="2025-10-01 09:01:24.190597212 +0000 UTC m=+6941.032204667" Oct 01 09:01:29 crc kubenswrapper[4837]: I1001 09:01:29.223337 4837 generic.go:334] "Generic (PLEG): container finished" podID="9ef08c36-8f34-4174-bc6f-d6bdaa96a498" containerID="8f5e9afc1bf928299fb9aeacbeb044a6916edd586d2dee2febe5daba3bb7575b" exitCode=0 Oct 01 09:01:29 crc kubenswrapper[4837]: I1001 09:01:29.223532 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbxg6" event={"ID":"9ef08c36-8f34-4174-bc6f-d6bdaa96a498","Type":"ContainerDied","Data":"8f5e9afc1bf928299fb9aeacbeb044a6916edd586d2dee2febe5daba3bb7575b"} Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.564344 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.672855 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data\") pod \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.673299 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89c5k\" (UniqueName: \"kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k\") pod \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.673440 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle\") pod \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\" (UID: \"9ef08c36-8f34-4174-bc6f-d6bdaa96a498\") " Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.704979 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k" (OuterVolumeSpecName: "kube-api-access-89c5k") pod "9ef08c36-8f34-4174-bc6f-d6bdaa96a498" (UID: "9ef08c36-8f34-4174-bc6f-d6bdaa96a498"). InnerVolumeSpecName "kube-api-access-89c5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.715909 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9ef08c36-8f34-4174-bc6f-d6bdaa96a498" (UID: "9ef08c36-8f34-4174-bc6f-d6bdaa96a498"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.766855 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ef08c36-8f34-4174-bc6f-d6bdaa96a498" (UID: "9ef08c36-8f34-4174-bc6f-d6bdaa96a498"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.777637 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89c5k\" (UniqueName: \"kubernetes.io/projected/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-kube-api-access-89c5k\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.777682 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:30 crc kubenswrapper[4837]: I1001 09:01:30.777713 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ef08c36-8f34-4174-bc6f-d6bdaa96a498-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.241000 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mbxg6" event={"ID":"9ef08c36-8f34-4174-bc6f-d6bdaa96a498","Type":"ContainerDied","Data":"7e8cc5e0f38b5e1918bf1754ce02e6c6148f54d64983df8264594e89052e01fc"} Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.241043 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e8cc5e0f38b5e1918bf1754ce02e6c6148f54d64983df8264594e89052e01fc" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.241104 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mbxg6" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.503052 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7bb6d5579f-w8v2b"] Oct 01 09:01:31 crc kubenswrapper[4837]: E1001 09:01:31.503667 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef08c36-8f34-4174-bc6f-d6bdaa96a498" containerName="barbican-db-sync" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.503688 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef08c36-8f34-4174-bc6f-d6bdaa96a498" containerName="barbican-db-sync" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.503887 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef08c36-8f34-4174-bc6f-d6bdaa96a498" containerName="barbican-db-sync" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.504787 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.506621 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dvvbm" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.509917 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.510119 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.530404 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7bb6d5579f-w8v2b"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.538011 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-69547f7db-nvq8j"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.542141 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.548026 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.579345 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69547f7db-nvq8j"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.593881 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/deb6c8d1-7a9e-4b84-b12e-6b744c500283-logs\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594194 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c89fw\" (UniqueName: \"kubernetes.io/projected/418f6100-faf8-46ae-b22d-bf013ff3c850-kube-api-access-c89fw\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594378 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-combined-ca-bundle\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594486 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594580 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594692 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-combined-ca-bundle\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data-custom\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.594960 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n42zq\" (UniqueName: \"kubernetes.io/projected/deb6c8d1-7a9e-4b84-b12e-6b744c500283-kube-api-access-n42zq\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.595141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/418f6100-faf8-46ae-b22d-bf013ff3c850-logs\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.595287 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data-custom\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.647754 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.653953 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.669461 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699106 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/deb6c8d1-7a9e-4b84-b12e-6b744c500283-logs\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699257 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c89fw\" (UniqueName: \"kubernetes.io/projected/418f6100-faf8-46ae-b22d-bf013ff3c850-kube-api-access-c89fw\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699293 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-combined-ca-bundle\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699333 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699352 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8mrs\" (UniqueName: \"kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699378 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699421 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-combined-ca-bundle\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699486 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data-custom\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699540 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n42zq\" (UniqueName: \"kubernetes.io/projected/deb6c8d1-7a9e-4b84-b12e-6b744c500283-kube-api-access-n42zq\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699593 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699667 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/418f6100-faf8-46ae-b22d-bf013ff3c850-logs\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699760 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data-custom\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.699797 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.702108 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/deb6c8d1-7a9e-4b84-b12e-6b744c500283-logs\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.702384 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/418f6100-faf8-46ae-b22d-bf013ff3c850-logs\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.706831 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.710038 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data-custom\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.710507 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-config-data\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.713309 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb6c8d1-7a9e-4b84-b12e-6b744c500283-combined-ca-bundle\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.736399 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-config-data-custom\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.736513 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418f6100-faf8-46ae-b22d-bf013ff3c850-combined-ca-bundle\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.740262 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n42zq\" (UniqueName: \"kubernetes.io/projected/deb6c8d1-7a9e-4b84-b12e-6b744c500283-kube-api-access-n42zq\") pod \"barbican-keystone-listener-69547f7db-nvq8j\" (UID: \"deb6c8d1-7a9e-4b84-b12e-6b744c500283\") " pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.746224 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c89fw\" (UniqueName: \"kubernetes.io/projected/418f6100-faf8-46ae-b22d-bf013ff3c850-kube-api-access-c89fw\") pod \"barbican-worker-7bb6d5579f-w8v2b\" (UID: \"418f6100-faf8-46ae-b22d-bf013ff3c850\") " pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.801532 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.801630 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.801660 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.801776 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.801796 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8mrs\" (UniqueName: \"kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.803097 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.803275 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.803337 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.803676 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.822362 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.878233 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8mrs\" (UniqueName: \"kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs\") pod \"dnsmasq-dns-67bcd8675-skwn2\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.888141 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.895512 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.897118 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.900302 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.908624 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:31 crc kubenswrapper[4837]: I1001 09:01:31.973671 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.007380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.007421 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6ms\" (UniqueName: \"kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.007544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.007572 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.007726 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.109451 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.109864 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6ms\" (UniqueName: \"kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.109980 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.110019 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.110044 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.110454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.116713 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.121399 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.123666 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.136805 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6ms\" (UniqueName: \"kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms\") pod \"barbican-api-556dccfd86-8djc6\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.259920 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.267034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69547f7db-nvq8j"] Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.335849 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7bb6d5579f-w8v2b"] Oct 01 09:01:32 crc kubenswrapper[4837]: W1001 09:01:32.344060 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418f6100_faf8_46ae_b22d_bf013ff3c850.slice/crio-945c8a960b5dc380707e37aac01feee84106798515cf2c467e9fa84c65ab38d7 WatchSource:0}: Error finding container 945c8a960b5dc380707e37aac01feee84106798515cf2c467e9fa84c65ab38d7: Status 404 returned error can't find the container with id 945c8a960b5dc380707e37aac01feee84106798515cf2c467e9fa84c65ab38d7 Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.570912 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:01:32 crc kubenswrapper[4837]: I1001 09:01:32.861851 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.271156 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" event={"ID":"deb6c8d1-7a9e-4b84-b12e-6b744c500283","Type":"ContainerStarted","Data":"b9e3f1917020620ffd4f1baed98facdf854659dd11a78b527efe8a5b47bd42e9"} Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.272742 4837 generic.go:334] "Generic (PLEG): container finished" podID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerID="6dece9d37f8a1e31b9b030470eed6ea3af9c2829f0a36cd0812ff7104d3ae375" exitCode=0 Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.272824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" event={"ID":"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c","Type":"ContainerDied","Data":"6dece9d37f8a1e31b9b030470eed6ea3af9c2829f0a36cd0812ff7104d3ae375"} Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.272854 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" event={"ID":"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c","Type":"ContainerStarted","Data":"59bb93f068801abdbf32a88127932b95faf66f20705edbd75cdf4610259b2f83"} Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.277913 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerStarted","Data":"a7e6a01814e97bc082b9fc944b216f076ff85d09615959748df782cfdeb58ce8"} Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.277959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerStarted","Data":"849048b58514c3b4d588da060293d7fc534ec7e662c5810c4729a9b1edaa41ab"} Oct 01 09:01:33 crc kubenswrapper[4837]: I1001 09:01:33.279382 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" event={"ID":"418f6100-faf8-46ae-b22d-bf013ff3c850","Type":"ContainerStarted","Data":"945c8a960b5dc380707e37aac01feee84106798515cf2c467e9fa84c65ab38d7"} Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.296885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerStarted","Data":"272b9a945880284fa57148276bd35a894bb79e12e8c981c437873fdc3ac5d178"} Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.298112 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.298220 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.324568 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-556dccfd86-8djc6" podStartSLOduration=3.3245429 podStartE2EDuration="3.3245429s" podCreationTimestamp="2025-10-01 09:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:01:34.315657911 +0000 UTC m=+6951.157265366" watchObservedRunningTime="2025-10-01 09:01:34.3245429 +0000 UTC m=+6951.166150355" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.662021 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54d786f8b6-vxnx7"] Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.668247 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.670931 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54d786f8b6-vxnx7"] Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.673167 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.673501 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.774920 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ece5705-7965-4476-ae5d-7ef6e80970dd-logs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775009 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-combined-ca-bundle\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775054 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-public-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775129 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfd7\" (UniqueName: \"kubernetes.io/projected/1ece5705-7965-4476-ae5d-7ef6e80970dd-kube-api-access-nvfd7\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775157 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-internal-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.775183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data-custom\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.876759 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ece5705-7965-4476-ae5d-7ef6e80970dd-logs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877133 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-combined-ca-bundle\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877420 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-public-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877525 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfd7\" (UniqueName: \"kubernetes.io/projected/1ece5705-7965-4476-ae5d-7ef6e80970dd-kube-api-access-nvfd7\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877645 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-internal-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877781 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data-custom\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.877300 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ece5705-7965-4476-ae5d-7ef6e80970dd-logs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.883626 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data-custom\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.884364 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-config-data\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.884450 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-public-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.885213 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-internal-tls-certs\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.886853 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ece5705-7965-4476-ae5d-7ef6e80970dd-combined-ca-bundle\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:34 crc kubenswrapper[4837]: I1001 09:01:34.900848 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfd7\" (UniqueName: \"kubernetes.io/projected/1ece5705-7965-4476-ae5d-7ef6e80970dd-kube-api-access-nvfd7\") pod \"barbican-api-54d786f8b6-vxnx7\" (UID: \"1ece5705-7965-4476-ae5d-7ef6e80970dd\") " pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.002309 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.315361 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" event={"ID":"418f6100-faf8-46ae-b22d-bf013ff3c850","Type":"ContainerStarted","Data":"0c8256cd15245df19a051779884b1760d71506d3c0bdca4b0135948fb1084df5"} Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.320264 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" event={"ID":"deb6c8d1-7a9e-4b84-b12e-6b744c500283","Type":"ContainerStarted","Data":"ba2ad9d3cc9ac51e716605b5aa66e6b16a913274779f2ec64330477548cd5c55"} Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.331291 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" event={"ID":"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c","Type":"ContainerStarted","Data":"caeae328ad87f649439b55b74f29c599c5af547745f919fcda9612dd5da5c1ab"} Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.331420 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.354344 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" podStartSLOduration=4.354321596 podStartE2EDuration="4.354321596s" podCreationTimestamp="2025-10-01 09:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:01:35.35242739 +0000 UTC m=+6952.194034845" watchObservedRunningTime="2025-10-01 09:01:35.354321596 +0000 UTC m=+6952.195929051" Oct 01 09:01:35 crc kubenswrapper[4837]: I1001 09:01:35.486103 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54d786f8b6-vxnx7"] Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.339742 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" event={"ID":"418f6100-faf8-46ae-b22d-bf013ff3c850","Type":"ContainerStarted","Data":"81d96d147c1c8d1c7bcca706c4ae8e56714863905cf93021a8eb840751ba6139"} Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.343765 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" event={"ID":"deb6c8d1-7a9e-4b84-b12e-6b744c500283","Type":"ContainerStarted","Data":"4943bfa45ce0d8e0961f64464ca451399efa422b37c32a4a03e1fd7374600a23"} Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.346133 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54d786f8b6-vxnx7" event={"ID":"1ece5705-7965-4476-ae5d-7ef6e80970dd","Type":"ContainerStarted","Data":"f50e83c5e03bd78338f4cd0e5fa5996dadb6083abb6aeda4e1257f777ad372cc"} Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.346200 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54d786f8b6-vxnx7" event={"ID":"1ece5705-7965-4476-ae5d-7ef6e80970dd","Type":"ContainerStarted","Data":"bc6e296932775d69b382d1baaadcfbffeb73eb565f4b5657466897e8f1a76547"} Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.346216 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54d786f8b6-vxnx7" event={"ID":"1ece5705-7965-4476-ae5d-7ef6e80970dd","Type":"ContainerStarted","Data":"e06f0852597e1c607dcbd3aa8652220793d8c7fa00730d5d9dd96f36f813be1e"} Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.346325 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.346381 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.366779 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7bb6d5579f-w8v2b" podStartSLOduration=2.989286424 podStartE2EDuration="5.366757616s" podCreationTimestamp="2025-10-01 09:01:31 +0000 UTC" firstStartedPulling="2025-10-01 09:01:32.353158113 +0000 UTC m=+6949.194765568" lastFinishedPulling="2025-10-01 09:01:34.730629315 +0000 UTC m=+6951.572236760" observedRunningTime="2025-10-01 09:01:36.361492478 +0000 UTC m=+6953.203099933" watchObservedRunningTime="2025-10-01 09:01:36.366757616 +0000 UTC m=+6953.208365071" Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.386677 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-69547f7db-nvq8j" podStartSLOduration=2.945090729 podStartE2EDuration="5.386656876s" podCreationTimestamp="2025-10-01 09:01:31 +0000 UTC" firstStartedPulling="2025-10-01 09:01:32.288982697 +0000 UTC m=+6949.130590142" lastFinishedPulling="2025-10-01 09:01:34.730548834 +0000 UTC m=+6951.572156289" observedRunningTime="2025-10-01 09:01:36.381876888 +0000 UTC m=+6953.223484343" watchObservedRunningTime="2025-10-01 09:01:36.386656876 +0000 UTC m=+6953.228264331" Oct 01 09:01:36 crc kubenswrapper[4837]: I1001 09:01:36.401835 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54d786f8b6-vxnx7" podStartSLOduration=2.401810717 podStartE2EDuration="2.401810717s" podCreationTimestamp="2025-10-01 09:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:01:36.398393694 +0000 UTC m=+6953.240001139" watchObservedRunningTime="2025-10-01 09:01:36.401810717 +0000 UTC m=+6953.243418162" Oct 01 09:01:40 crc kubenswrapper[4837]: I1001 09:01:40.108055 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.573069 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.605456 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54d786f8b6-vxnx7" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.685925 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.688189 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" containerID="cri-o://a7e6a01814e97bc082b9fc944b216f076ff85d09615959748df782cfdeb58ce8" gracePeriod=30 Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.688888 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" containerID="cri-o://272b9a945880284fa57148276bd35a894bb79e12e8c981c437873fdc3ac5d178" gracePeriod=30 Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.696974 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": EOF" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.697021 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": EOF" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.697043 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": EOF" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.697163 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": EOF" Oct 01 09:01:41 crc kubenswrapper[4837]: I1001 09:01:41.976510 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.073953 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.074205 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="dnsmasq-dns" containerID="cri-o://6cdacb8788bab3e066e8a50ad5011bc7b7cdee4430d7adae479cc687b136d1dd" gracePeriod=10 Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.400861 4837 generic.go:334] "Generic (PLEG): container finished" podID="4538c208-8c4d-4565-aa65-53cf2b137691" containerID="6cdacb8788bab3e066e8a50ad5011bc7b7cdee4430d7adae479cc687b136d1dd" exitCode=0 Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.400959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" event={"ID":"4538c208-8c4d-4565-aa65-53cf2b137691","Type":"ContainerDied","Data":"6cdacb8788bab3e066e8a50ad5011bc7b7cdee4430d7adae479cc687b136d1dd"} Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.403373 4837 generic.go:334] "Generic (PLEG): container finished" podID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerID="a7e6a01814e97bc082b9fc944b216f076ff85d09615959748df782cfdeb58ce8" exitCode=143 Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.403448 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerDied","Data":"a7e6a01814e97bc082b9fc944b216f076ff85d09615959748df782cfdeb58ce8"} Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.657231 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.841683 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k\") pod \"4538c208-8c4d-4565-aa65-53cf2b137691\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.841770 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config\") pod \"4538c208-8c4d-4565-aa65-53cf2b137691\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.841849 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc\") pod \"4538c208-8c4d-4565-aa65-53cf2b137691\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.841885 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb\") pod \"4538c208-8c4d-4565-aa65-53cf2b137691\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.841936 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb\") pod \"4538c208-8c4d-4565-aa65-53cf2b137691\" (UID: \"4538c208-8c4d-4565-aa65-53cf2b137691\") " Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.886662 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k" (OuterVolumeSpecName: "kube-api-access-mx86k") pod "4538c208-8c4d-4565-aa65-53cf2b137691" (UID: "4538c208-8c4d-4565-aa65-53cf2b137691"). InnerVolumeSpecName "kube-api-access-mx86k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.919253 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config" (OuterVolumeSpecName: "config") pod "4538c208-8c4d-4565-aa65-53cf2b137691" (UID: "4538c208-8c4d-4565-aa65-53cf2b137691"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.945019 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.945058 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx86k\" (UniqueName: \"kubernetes.io/projected/4538c208-8c4d-4565-aa65-53cf2b137691-kube-api-access-mx86k\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.947841 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4538c208-8c4d-4565-aa65-53cf2b137691" (UID: "4538c208-8c4d-4565-aa65-53cf2b137691"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:01:42 crc kubenswrapper[4837]: I1001 09:01:42.972536 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4538c208-8c4d-4565-aa65-53cf2b137691" (UID: "4538c208-8c4d-4565-aa65-53cf2b137691"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.002609 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4538c208-8c4d-4565-aa65-53cf2b137691" (UID: "4538c208-8c4d-4565-aa65-53cf2b137691"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.061655 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.061877 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.062008 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4538c208-8c4d-4565-aa65-53cf2b137691-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.429856 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" event={"ID":"4538c208-8c4d-4565-aa65-53cf2b137691","Type":"ContainerDied","Data":"c994fef84f2912789316a0c9752aa28826a3af4590153f550a01c8a74272924c"} Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.429937 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9c787459-zc4c9" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.429940 4837 scope.go:117] "RemoveContainer" containerID="6cdacb8788bab3e066e8a50ad5011bc7b7cdee4430d7adae479cc687b136d1dd" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.456095 4837 scope.go:117] "RemoveContainer" containerID="a6e667d6ec6f4c61201bcce7746615daf9311ef8391ec4fcaf710f1efd9bc58a" Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.466987 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.474315 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f9c787459-zc4c9"] Oct 01 09:01:43 crc kubenswrapper[4837]: I1001 09:01:43.827885 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" path="/var/lib/kubelet/pods/4538c208-8c4d-4565-aa65-53cf2b137691/volumes" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.100709 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": read tcp 10.217.0.2:48008->10.217.1.45:9311: read: connection reset by peer" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.100661 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": read tcp 10.217.0.2:48002->10.217.1.45:9311: read: connection reset by peer" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.261038 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": dial tcp 10.217.1.45:9311: connect: connection refused" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.261075 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-556dccfd86-8djc6" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.45:9311/healthcheck\": dial tcp 10.217.1.45:9311: connect: connection refused" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.475212 4837 generic.go:334] "Generic (PLEG): container finished" podID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerID="272b9a945880284fa57148276bd35a894bb79e12e8c981c437873fdc3ac5d178" exitCode=0 Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.475321 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerDied","Data":"272b9a945880284fa57148276bd35a894bb79e12e8c981c437873fdc3ac5d178"} Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.615283 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.666891 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom\") pod \"2fec7f82-5447-441e-90ba-c8539e1e2aae\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.666996 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs\") pod \"2fec7f82-5447-441e-90ba-c8539e1e2aae\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.667102 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle\") pod \"2fec7f82-5447-441e-90ba-c8539e1e2aae\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.667229 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data\") pod \"2fec7f82-5447-441e-90ba-c8539e1e2aae\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.667283 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf6ms\" (UniqueName: \"kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms\") pod \"2fec7f82-5447-441e-90ba-c8539e1e2aae\" (UID: \"2fec7f82-5447-441e-90ba-c8539e1e2aae\") " Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.668392 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs" (OuterVolumeSpecName: "logs") pod "2fec7f82-5447-441e-90ba-c8539e1e2aae" (UID: "2fec7f82-5447-441e-90ba-c8539e1e2aae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.692010 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms" (OuterVolumeSpecName: "kube-api-access-rf6ms") pod "2fec7f82-5447-441e-90ba-c8539e1e2aae" (UID: "2fec7f82-5447-441e-90ba-c8539e1e2aae"). InnerVolumeSpecName "kube-api-access-rf6ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.692149 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2fec7f82-5447-441e-90ba-c8539e1e2aae" (UID: "2fec7f82-5447-441e-90ba-c8539e1e2aae"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.756009 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fec7f82-5447-441e-90ba-c8539e1e2aae" (UID: "2fec7f82-5447-441e-90ba-c8539e1e2aae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.770843 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.770887 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fec7f82-5447-441e-90ba-c8539e1e2aae-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.770898 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.770909 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf6ms\" (UniqueName: \"kubernetes.io/projected/2fec7f82-5447-441e-90ba-c8539e1e2aae-kube-api-access-rf6ms\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.835783 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data" (OuterVolumeSpecName: "config-data") pod "2fec7f82-5447-441e-90ba-c8539e1e2aae" (UID: "2fec7f82-5447-441e-90ba-c8539e1e2aae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:01:47 crc kubenswrapper[4837]: I1001 09:01:47.873240 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fec7f82-5447-441e-90ba-c8539e1e2aae-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.488438 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-556dccfd86-8djc6" event={"ID":"2fec7f82-5447-441e-90ba-c8539e1e2aae","Type":"ContainerDied","Data":"849048b58514c3b4d588da060293d7fc534ec7e662c5810c4729a9b1edaa41ab"} Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.488536 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-556dccfd86-8djc6" Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.488853 4837 scope.go:117] "RemoveContainer" containerID="272b9a945880284fa57148276bd35a894bb79e12e8c981c437873fdc3ac5d178" Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.518236 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.525450 4837 scope.go:117] "RemoveContainer" containerID="a7e6a01814e97bc082b9fc944b216f076ff85d09615959748df782cfdeb58ce8" Oct 01 09:01:48 crc kubenswrapper[4837]: I1001 09:01:48.527741 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-556dccfd86-8djc6"] Oct 01 09:01:49 crc kubenswrapper[4837]: I1001 09:01:49.842606 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" path="/var/lib/kubelet/pods/2fec7f82-5447-441e-90ba-c8539e1e2aae/volumes" Oct 01 09:01:53 crc kubenswrapper[4837]: I1001 09:01:53.080004 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:01:53 crc kubenswrapper[4837]: I1001 09:01:53.080296 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.039297 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-dl2w4"] Oct 01 09:02:13 crc kubenswrapper[4837]: E1001 09:02:13.040203 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040234 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" Oct 01 09:02:13 crc kubenswrapper[4837]: E1001 09:02:13.040270 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="init" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040276 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="init" Oct 01 09:02:13 crc kubenswrapper[4837]: E1001 09:02:13.040286 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="dnsmasq-dns" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040292 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="dnsmasq-dns" Oct 01 09:02:13 crc kubenswrapper[4837]: E1001 09:02:13.040305 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040310 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040461 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040484 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4538c208-8c4d-4565-aa65-53cf2b137691" containerName="dnsmasq-dns" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.040491 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fec7f82-5447-441e-90ba-c8539e1e2aae" containerName="barbican-api-log" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.041091 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.051538 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dl2w4"] Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.072082 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bsw\" (UniqueName: \"kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw\") pod \"neutron-db-create-dl2w4\" (UID: \"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c\") " pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.173724 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bsw\" (UniqueName: \"kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw\") pod \"neutron-db-create-dl2w4\" (UID: \"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c\") " pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.198286 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bsw\" (UniqueName: \"kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw\") pod \"neutron-db-create-dl2w4\" (UID: \"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c\") " pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:13 crc kubenswrapper[4837]: I1001 09:02:13.371024 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:14 crc kubenswrapper[4837]: I1001 09:02:14.510811 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dl2w4"] Oct 01 09:02:14 crc kubenswrapper[4837]: I1001 09:02:14.740738 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dl2w4" event={"ID":"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c","Type":"ContainerStarted","Data":"e653f71333e1d55c71b277b79feb0e810a2ff196e85a96d4baded6dae9104645"} Oct 01 09:02:14 crc kubenswrapper[4837]: I1001 09:02:14.740800 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dl2w4" event={"ID":"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c","Type":"ContainerStarted","Data":"5eccb49f527920a5f5c82e796e4aa292b1655ff732378420b7612b9d2e07dc52"} Oct 01 09:02:15 crc kubenswrapper[4837]: I1001 09:02:15.754148 4837 generic.go:334] "Generic (PLEG): container finished" podID="bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" containerID="e653f71333e1d55c71b277b79feb0e810a2ff196e85a96d4baded6dae9104645" exitCode=0 Oct 01 09:02:15 crc kubenswrapper[4837]: I1001 09:02:15.754644 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dl2w4" event={"ID":"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c","Type":"ContainerDied","Data":"e653f71333e1d55c71b277b79feb0e810a2ff196e85a96d4baded6dae9104645"} Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.119216 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.127741 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5bsw\" (UniqueName: \"kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw\") pod \"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c\" (UID: \"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c\") " Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.140914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw" (OuterVolumeSpecName: "kube-api-access-k5bsw") pod "bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" (UID: "bcaeed1f-08d9-4633-bc53-c0f2cb10f38c"). InnerVolumeSpecName "kube-api-access-k5bsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.229642 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5bsw\" (UniqueName: \"kubernetes.io/projected/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c-kube-api-access-k5bsw\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.766867 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dl2w4" event={"ID":"bcaeed1f-08d9-4633-bc53-c0f2cb10f38c","Type":"ContainerDied","Data":"5eccb49f527920a5f5c82e796e4aa292b1655ff732378420b7612b9d2e07dc52"} Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.766924 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eccb49f527920a5f5c82e796e4aa292b1655ff732378420b7612b9d2e07dc52" Oct 01 09:02:16 crc kubenswrapper[4837]: I1001 09:02:16.766995 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dl2w4" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.080418 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.081162 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.160960 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b796-account-create-7g2r6"] Oct 01 09:02:23 crc kubenswrapper[4837]: E1001 09:02:23.161863 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" containerName="mariadb-database-create" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.161911 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" containerName="mariadb-database-create" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.162425 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" containerName="mariadb-database-create" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.164023 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.167362 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.169095 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b796-account-create-7g2r6"] Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.272363 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq6wq\" (UniqueName: \"kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq\") pod \"neutron-b796-account-create-7g2r6\" (UID: \"052fcef6-6d55-4cb1-9b9a-111a49189b31\") " pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.374874 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq6wq\" (UniqueName: \"kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq\") pod \"neutron-b796-account-create-7g2r6\" (UID: \"052fcef6-6d55-4cb1-9b9a-111a49189b31\") " pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.398765 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq6wq\" (UniqueName: \"kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq\") pod \"neutron-b796-account-create-7g2r6\" (UID: \"052fcef6-6d55-4cb1-9b9a-111a49189b31\") " pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.486100 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:23 crc kubenswrapper[4837]: I1001 09:02:23.949588 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b796-account-create-7g2r6"] Oct 01 09:02:24 crc kubenswrapper[4837]: I1001 09:02:24.840442 4837 generic.go:334] "Generic (PLEG): container finished" podID="052fcef6-6d55-4cb1-9b9a-111a49189b31" containerID="ebf9bb932aa70cb2e5313374790bf84e50d8593ca7e4969f1cc332ce43c99342" exitCode=0 Oct 01 09:02:24 crc kubenswrapper[4837]: I1001 09:02:24.840513 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b796-account-create-7g2r6" event={"ID":"052fcef6-6d55-4cb1-9b9a-111a49189b31","Type":"ContainerDied","Data":"ebf9bb932aa70cb2e5313374790bf84e50d8593ca7e4969f1cc332ce43c99342"} Oct 01 09:02:24 crc kubenswrapper[4837]: I1001 09:02:24.840925 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b796-account-create-7g2r6" event={"ID":"052fcef6-6d55-4cb1-9b9a-111a49189b31","Type":"ContainerStarted","Data":"57b2cacf8d8ffc08c69dfa1ee4d1bc2d629b156e434c518350d9e48c4b6c45a5"} Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.236421 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.340017 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq6wq\" (UniqueName: \"kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq\") pod \"052fcef6-6d55-4cb1-9b9a-111a49189b31\" (UID: \"052fcef6-6d55-4cb1-9b9a-111a49189b31\") " Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.346906 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq" (OuterVolumeSpecName: "kube-api-access-mq6wq") pod "052fcef6-6d55-4cb1-9b9a-111a49189b31" (UID: "052fcef6-6d55-4cb1-9b9a-111a49189b31"). InnerVolumeSpecName "kube-api-access-mq6wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.442657 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq6wq\" (UniqueName: \"kubernetes.io/projected/052fcef6-6d55-4cb1-9b9a-111a49189b31-kube-api-access-mq6wq\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.866444 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b796-account-create-7g2r6" event={"ID":"052fcef6-6d55-4cb1-9b9a-111a49189b31","Type":"ContainerDied","Data":"57b2cacf8d8ffc08c69dfa1ee4d1bc2d629b156e434c518350d9e48c4b6c45a5"} Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.866506 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57b2cacf8d8ffc08c69dfa1ee4d1bc2d629b156e434c518350d9e48c4b6c45a5" Oct 01 09:02:26 crc kubenswrapper[4837]: I1001 09:02:26.866552 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b796-account-create-7g2r6" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.332982 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wlkmb"] Oct 01 09:02:28 crc kubenswrapper[4837]: E1001 09:02:28.333880 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052fcef6-6d55-4cb1-9b9a-111a49189b31" containerName="mariadb-account-create" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.333899 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="052fcef6-6d55-4cb1-9b9a-111a49189b31" containerName="mariadb-account-create" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.334156 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="052fcef6-6d55-4cb1-9b9a-111a49189b31" containerName="mariadb-account-create" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.334938 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.337983 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.338273 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.339617 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j8sg4" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.342741 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wlkmb"] Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.483079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.483179 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.483210 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w9tb\" (UniqueName: \"kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.585388 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.585529 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.585601 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w9tb\" (UniqueName: \"kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.594981 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.598269 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.621792 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w9tb\" (UniqueName: \"kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb\") pod \"neutron-db-sync-wlkmb\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.653132 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:28 crc kubenswrapper[4837]: I1001 09:02:28.922856 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wlkmb"] Oct 01 09:02:29 crc kubenswrapper[4837]: I1001 09:02:29.894422 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wlkmb" event={"ID":"36fcdeac-e848-4460-93e2-ea6173928012","Type":"ContainerStarted","Data":"d41c236862bbb1755f8ba5864b9eca5ca1bd474ab6b95aebbf365a136c0b1559"} Oct 01 09:02:29 crc kubenswrapper[4837]: I1001 09:02:29.894914 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wlkmb" event={"ID":"36fcdeac-e848-4460-93e2-ea6173928012","Type":"ContainerStarted","Data":"a657b6e53c513405129d8f4180c1373632cfa72dffc07a4de894fea213fb1d22"} Oct 01 09:02:29 crc kubenswrapper[4837]: I1001 09:02:29.917033 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wlkmb" podStartSLOduration=1.916999079 podStartE2EDuration="1.916999079s" podCreationTimestamp="2025-10-01 09:02:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:02:29.911346391 +0000 UTC m=+7006.752953856" watchObservedRunningTime="2025-10-01 09:02:29.916999079 +0000 UTC m=+7006.758606534" Oct 01 09:02:36 crc kubenswrapper[4837]: I1001 09:02:36.985877 4837 generic.go:334] "Generic (PLEG): container finished" podID="36fcdeac-e848-4460-93e2-ea6173928012" containerID="d41c236862bbb1755f8ba5864b9eca5ca1bd474ab6b95aebbf365a136c0b1559" exitCode=0 Oct 01 09:02:36 crc kubenswrapper[4837]: I1001 09:02:36.985952 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wlkmb" event={"ID":"36fcdeac-e848-4460-93e2-ea6173928012","Type":"ContainerDied","Data":"d41c236862bbb1755f8ba5864b9eca5ca1bd474ab6b95aebbf365a136c0b1559"} Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.378136 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.484515 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config\") pod \"36fcdeac-e848-4460-93e2-ea6173928012\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.484756 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w9tb\" (UniqueName: \"kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb\") pod \"36fcdeac-e848-4460-93e2-ea6173928012\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.484862 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle\") pod \"36fcdeac-e848-4460-93e2-ea6173928012\" (UID: \"36fcdeac-e848-4460-93e2-ea6173928012\") " Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.494203 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb" (OuterVolumeSpecName: "kube-api-access-8w9tb") pod "36fcdeac-e848-4460-93e2-ea6173928012" (UID: "36fcdeac-e848-4460-93e2-ea6173928012"). InnerVolumeSpecName "kube-api-access-8w9tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.514089 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36fcdeac-e848-4460-93e2-ea6173928012" (UID: "36fcdeac-e848-4460-93e2-ea6173928012"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.527275 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config" (OuterVolumeSpecName: "config") pod "36fcdeac-e848-4460-93e2-ea6173928012" (UID: "36fcdeac-e848-4460-93e2-ea6173928012"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.587147 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w9tb\" (UniqueName: \"kubernetes.io/projected/36fcdeac-e848-4460-93e2-ea6173928012-kube-api-access-8w9tb\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.587190 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:38 crc kubenswrapper[4837]: I1001 09:02:38.587203 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/36fcdeac-e848-4460-93e2-ea6173928012-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.026137 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wlkmb" event={"ID":"36fcdeac-e848-4460-93e2-ea6173928012","Type":"ContainerDied","Data":"a657b6e53c513405129d8f4180c1373632cfa72dffc07a4de894fea213fb1d22"} Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.026595 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a657b6e53c513405129d8f4180c1373632cfa72dffc07a4de894fea213fb1d22" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.026460 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wlkmb" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.281388 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:02:39 crc kubenswrapper[4837]: E1001 09:02:39.281934 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36fcdeac-e848-4460-93e2-ea6173928012" containerName="neutron-db-sync" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.281962 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="36fcdeac-e848-4460-93e2-ea6173928012" containerName="neutron-db-sync" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.282204 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="36fcdeac-e848-4460-93e2-ea6173928012" containerName="neutron-db-sync" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.283638 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.297442 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.367707 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.369631 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.376185 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j8sg4" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.376425 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.376590 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.381013 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.385447 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.404275 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.404332 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xmph\" (UniqueName: \"kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.404389 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.404425 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.404460 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506169 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506252 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506285 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xmph\" (UniqueName: \"kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506320 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxdcl\" (UniqueName: \"kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506384 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506411 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506458 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506509 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506530 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.506555 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.507665 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.508074 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.508303 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.508453 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.536521 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xmph\" (UniqueName: \"kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph\") pod \"dnsmasq-dns-869fcf5fcf-kwxkg\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.608952 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.609469 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.609547 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.609570 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.609627 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.609654 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxdcl\" (UniqueName: \"kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.614605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.614933 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.617391 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.617921 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.637866 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxdcl\" (UniqueName: \"kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl\") pod \"neutron-7b86f4c74d-dlwfm\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:39 crc kubenswrapper[4837]: I1001 09:02:39.685132 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:40 crc kubenswrapper[4837]: I1001 09:02:40.187388 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:02:40 crc kubenswrapper[4837]: I1001 09:02:40.374733 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:02:40 crc kubenswrapper[4837]: W1001 09:02:40.376530 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod719662c0_7881_465a_b1ab_dffb9b07f386.slice/crio-842cd4abb047404a4e09bec841bd7419a5aadf44611e2a7b055a56c446f7380f WatchSource:0}: Error finding container 842cd4abb047404a4e09bec841bd7419a5aadf44611e2a7b055a56c446f7380f: Status 404 returned error can't find the container with id 842cd4abb047404a4e09bec841bd7419a5aadf44611e2a7b055a56c446f7380f Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.047836 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerStarted","Data":"a97dc4c27b2e2b2e6a0562314913dcab48f80375c45843b3d759964f3626ff24"} Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.049550 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerStarted","Data":"51e03baaed6c786567a2893a03a2985205febb5ae27ccd93507bbd1ea31349ed"} Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.049643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerStarted","Data":"842cd4abb047404a4e09bec841bd7419a5aadf44611e2a7b055a56c446f7380f"} Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.050205 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.053081 4837 generic.go:334] "Generic (PLEG): container finished" podID="3b19b130-2f32-40fe-afac-80cae13880f7" containerID="25d709f52e8b9e13ce453c52a7135b37f25934800aa0583b8ec5b2cc0a17f4aa" exitCode=0 Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.053143 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" event={"ID":"3b19b130-2f32-40fe-afac-80cae13880f7","Type":"ContainerDied","Data":"25d709f52e8b9e13ce453c52a7135b37f25934800aa0583b8ec5b2cc0a17f4aa"} Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.053180 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" event={"ID":"3b19b130-2f32-40fe-afac-80cae13880f7","Type":"ContainerStarted","Data":"17f79d6ead13974668bb90313a10024fe27cecda7ff029bc00f36b2f13493671"} Oct 01 09:02:41 crc kubenswrapper[4837]: I1001 09:02:41.091195 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b86f4c74d-dlwfm" podStartSLOduration=2.09116827 podStartE2EDuration="2.09116827s" podCreationTimestamp="2025-10-01 09:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:02:41.081949914 +0000 UTC m=+7017.923557379" watchObservedRunningTime="2025-10-01 09:02:41.09116827 +0000 UTC m=+7017.932775735" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.071923 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" event={"ID":"3b19b130-2f32-40fe-afac-80cae13880f7","Type":"ContainerStarted","Data":"26289cb6f00e22a8b002ea5b7e6fb3e1b300f8d23fa798f2de0896520262a3a7"} Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.072144 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.099952 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" podStartSLOduration=3.0999335 podStartE2EDuration="3.0999335s" podCreationTimestamp="2025-10-01 09:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:02:42.094732982 +0000 UTC m=+7018.936340447" watchObservedRunningTime="2025-10-01 09:02:42.0999335 +0000 UTC m=+7018.941540955" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.424105 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bd98cb97c-4l4pg"] Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.426226 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.442650 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.442782 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.447498 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd98cb97c-4l4pg"] Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.567811 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-internal-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.567893 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-public-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.567932 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.567948 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-httpd-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.567994 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-ovndb-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.568014 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7285\" (UniqueName: \"kubernetes.io/projected/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-kube-api-access-p7285\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.568035 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-combined-ca-bundle\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670142 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-public-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670213 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-httpd-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670294 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-ovndb-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670516 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7285\" (UniqueName: \"kubernetes.io/projected/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-kube-api-access-p7285\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670662 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-combined-ca-bundle\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.670941 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-internal-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.691429 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.691644 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-public-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.691662 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-combined-ca-bundle\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.694924 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-internal-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.695012 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-ovndb-tls-certs\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.703790 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7285\" (UniqueName: \"kubernetes.io/projected/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-kube-api-access-p7285\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.713560 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c629d2d6-9a91-4a51-9504-1524ee0bc8ed-httpd-config\") pod \"neutron-bd98cb97c-4l4pg\" (UID: \"c629d2d6-9a91-4a51-9504-1524ee0bc8ed\") " pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:42 crc kubenswrapper[4837]: I1001 09:02:42.754582 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:43 crc kubenswrapper[4837]: I1001 09:02:43.383237 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd98cb97c-4l4pg"] Oct 01 09:02:43 crc kubenswrapper[4837]: W1001 09:02:43.386363 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc629d2d6_9a91_4a51_9504_1524ee0bc8ed.slice/crio-1d29271f7b5a501a99d796b04a14a9ff9dce02a490bd41015a8d9e171d883de1 WatchSource:0}: Error finding container 1d29271f7b5a501a99d796b04a14a9ff9dce02a490bd41015a8d9e171d883de1: Status 404 returned error can't find the container with id 1d29271f7b5a501a99d796b04a14a9ff9dce02a490bd41015a8d9e171d883de1 Oct 01 09:02:44 crc kubenswrapper[4837]: I1001 09:02:44.093962 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd98cb97c-4l4pg" event={"ID":"c629d2d6-9a91-4a51-9504-1524ee0bc8ed","Type":"ContainerStarted","Data":"5f2de15e2bff4af2f4e87f9e92e60463e21b885eeada07e4e702f6ca5b4164e8"} Oct 01 09:02:44 crc kubenswrapper[4837]: I1001 09:02:44.094585 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd98cb97c-4l4pg" event={"ID":"c629d2d6-9a91-4a51-9504-1524ee0bc8ed","Type":"ContainerStarted","Data":"244633598cdf243e50652f9212674e227699eff9047e208c3fcf7c29f3370ba3"} Oct 01 09:02:44 crc kubenswrapper[4837]: I1001 09:02:44.094614 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:02:44 crc kubenswrapper[4837]: I1001 09:02:44.094633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd98cb97c-4l4pg" event={"ID":"c629d2d6-9a91-4a51-9504-1524ee0bc8ed","Type":"ContainerStarted","Data":"1d29271f7b5a501a99d796b04a14a9ff9dce02a490bd41015a8d9e171d883de1"} Oct 01 09:02:44 crc kubenswrapper[4837]: I1001 09:02:44.125027 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bd98cb97c-4l4pg" podStartSLOduration=2.124987475 podStartE2EDuration="2.124987475s" podCreationTimestamp="2025-10-01 09:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:02:44.122787612 +0000 UTC m=+7020.964395067" watchObservedRunningTime="2025-10-01 09:02:44.124987475 +0000 UTC m=+7020.966594930" Oct 01 09:02:49 crc kubenswrapper[4837]: I1001 09:02:49.610936 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:02:49 crc kubenswrapper[4837]: I1001 09:02:49.693589 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:02:49 crc kubenswrapper[4837]: I1001 09:02:49.693866 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="dnsmasq-dns" containerID="cri-o://caeae328ad87f649439b55b74f29c599c5af547745f919fcda9612dd5da5c1ab" gracePeriod=10 Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.156811 4837 generic.go:334] "Generic (PLEG): container finished" podID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerID="caeae328ad87f649439b55b74f29c599c5af547745f919fcda9612dd5da5c1ab" exitCode=0 Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.156932 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" event={"ID":"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c","Type":"ContainerDied","Data":"caeae328ad87f649439b55b74f29c599c5af547745f919fcda9612dd5da5c1ab"} Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.157359 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" event={"ID":"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c","Type":"ContainerDied","Data":"59bb93f068801abdbf32a88127932b95faf66f20705edbd75cdf4610259b2f83"} Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.157382 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59bb93f068801abdbf32a88127932b95faf66f20705edbd75cdf4610259b2f83" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.176324 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.232631 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config\") pod \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.232747 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb\") pod \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.232946 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb\") pod \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.234097 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8mrs\" (UniqueName: \"kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs\") pod \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.234196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc\") pod \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\" (UID: \"6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c\") " Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.245372 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs" (OuterVolumeSpecName: "kube-api-access-z8mrs") pod "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" (UID: "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c"). InnerVolumeSpecName "kube-api-access-z8mrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.287075 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config" (OuterVolumeSpecName: "config") pod "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" (UID: "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.304449 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" (UID: "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.315101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" (UID: "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.317385 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" (UID: "6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.337575 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8mrs\" (UniqueName: \"kubernetes.io/projected/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-kube-api-access-z8mrs\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.337644 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.337659 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.337678 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:50 crc kubenswrapper[4837]: I1001 09:02:50.337709 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:02:51 crc kubenswrapper[4837]: I1001 09:02:51.164048 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcd8675-skwn2" Oct 01 09:02:51 crc kubenswrapper[4837]: I1001 09:02:51.203974 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:02:51 crc kubenswrapper[4837]: I1001 09:02:51.215475 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bcd8675-skwn2"] Oct 01 09:02:51 crc kubenswrapper[4837]: I1001 09:02:51.828200 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" path="/var/lib/kubelet/pods/6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c/volumes" Oct 01 09:02:53 crc kubenswrapper[4837]: I1001 09:02:53.079798 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:02:53 crc kubenswrapper[4837]: I1001 09:02:53.080426 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:02:53 crc kubenswrapper[4837]: I1001 09:02:53.081220 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:02:53 crc kubenswrapper[4837]: I1001 09:02:53.082394 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:02:53 crc kubenswrapper[4837]: I1001 09:02:53.082522 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" gracePeriod=600 Oct 01 09:02:53 crc kubenswrapper[4837]: E1001 09:02:53.223960 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:02:54 crc kubenswrapper[4837]: I1001 09:02:54.217592 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" exitCode=0 Oct 01 09:02:54 crc kubenswrapper[4837]: I1001 09:02:54.217900 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6"} Oct 01 09:02:54 crc kubenswrapper[4837]: I1001 09:02:54.218031 4837 scope.go:117] "RemoveContainer" containerID="68c3ef1dd26fcef4ce07e36d238bc63d8934c377dc334d0da9f6ad58c12053d5" Oct 01 09:02:54 crc kubenswrapper[4837]: I1001 09:02:54.219218 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:02:54 crc kubenswrapper[4837]: E1001 09:02:54.219638 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:05 crc kubenswrapper[4837]: I1001 09:03:05.816552 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:03:05 crc kubenswrapper[4837]: E1001 09:03:05.817486 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:06 crc kubenswrapper[4837]: I1001 09:03:06.690539 4837 scope.go:117] "RemoveContainer" containerID="980788b559a08bbfe5b603cd77525c34f1e8c9892a3a0ed4b54a305dea3dd012" Oct 01 09:03:09 crc kubenswrapper[4837]: I1001 09:03:09.693734 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:03:12 crc kubenswrapper[4837]: I1001 09:03:12.770269 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bd98cb97c-4l4pg" Oct 01 09:03:12 crc kubenswrapper[4837]: I1001 09:03:12.852989 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:03:12 crc kubenswrapper[4837]: I1001 09:03:12.862850 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b86f4c74d-dlwfm" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-api" containerID="cri-o://51e03baaed6c786567a2893a03a2985205febb5ae27ccd93507bbd1ea31349ed" gracePeriod=30 Oct 01 09:03:12 crc kubenswrapper[4837]: I1001 09:03:12.862919 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b86f4c74d-dlwfm" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-httpd" containerID="cri-o://a97dc4c27b2e2b2e6a0562314913dcab48f80375c45843b3d759964f3626ff24" gracePeriod=30 Oct 01 09:03:13 crc kubenswrapper[4837]: I1001 09:03:13.417792 4837 generic.go:334] "Generic (PLEG): container finished" podID="719662c0-7881-465a-b1ab-dffb9b07f386" containerID="a97dc4c27b2e2b2e6a0562314913dcab48f80375c45843b3d759964f3626ff24" exitCode=0 Oct 01 09:03:13 crc kubenswrapper[4837]: I1001 09:03:13.417835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerDied","Data":"a97dc4c27b2e2b2e6a0562314913dcab48f80375c45843b3d759964f3626ff24"} Oct 01 09:03:17 crc kubenswrapper[4837]: I1001 09:03:17.459031 4837 generic.go:334] "Generic (PLEG): container finished" podID="719662c0-7881-465a-b1ab-dffb9b07f386" containerID="51e03baaed6c786567a2893a03a2985205febb5ae27ccd93507bbd1ea31349ed" exitCode=0 Oct 01 09:03:17 crc kubenswrapper[4837]: I1001 09:03:17.459143 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerDied","Data":"51e03baaed6c786567a2893a03a2985205febb5ae27ccd93507bbd1ea31349ed"} Oct 01 09:03:17 crc kubenswrapper[4837]: I1001 09:03:17.909927 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.004046 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs\") pod \"719662c0-7881-465a-b1ab-dffb9b07f386\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.004149 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle\") pod \"719662c0-7881-465a-b1ab-dffb9b07f386\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.004422 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxdcl\" (UniqueName: \"kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl\") pod \"719662c0-7881-465a-b1ab-dffb9b07f386\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.004470 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config\") pod \"719662c0-7881-465a-b1ab-dffb9b07f386\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.004510 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config\") pod \"719662c0-7881-465a-b1ab-dffb9b07f386\" (UID: \"719662c0-7881-465a-b1ab-dffb9b07f386\") " Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.013312 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "719662c0-7881-465a-b1ab-dffb9b07f386" (UID: "719662c0-7881-465a-b1ab-dffb9b07f386"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.013411 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl" (OuterVolumeSpecName: "kube-api-access-qxdcl") pod "719662c0-7881-465a-b1ab-dffb9b07f386" (UID: "719662c0-7881-465a-b1ab-dffb9b07f386"). InnerVolumeSpecName "kube-api-access-qxdcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.064155 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config" (OuterVolumeSpecName: "config") pod "719662c0-7881-465a-b1ab-dffb9b07f386" (UID: "719662c0-7881-465a-b1ab-dffb9b07f386"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.074417 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "719662c0-7881-465a-b1ab-dffb9b07f386" (UID: "719662c0-7881-465a-b1ab-dffb9b07f386"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.075729 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "719662c0-7881-465a-b1ab-dffb9b07f386" (UID: "719662c0-7881-465a-b1ab-dffb9b07f386"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.107111 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxdcl\" (UniqueName: \"kubernetes.io/projected/719662c0-7881-465a-b1ab-dffb9b07f386-kube-api-access-qxdcl\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.107315 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.107414 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.108257 4837 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.108350 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/719662c0-7881-465a-b1ab-dffb9b07f386-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.473034 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b86f4c74d-dlwfm" event={"ID":"719662c0-7881-465a-b1ab-dffb9b07f386","Type":"ContainerDied","Data":"842cd4abb047404a4e09bec841bd7419a5aadf44611e2a7b055a56c446f7380f"} Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.474106 4837 scope.go:117] "RemoveContainer" containerID="a97dc4c27b2e2b2e6a0562314913dcab48f80375c45843b3d759964f3626ff24" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.474316 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b86f4c74d-dlwfm" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.518349 4837 scope.go:117] "RemoveContainer" containerID="51e03baaed6c786567a2893a03a2985205febb5ae27ccd93507bbd1ea31349ed" Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.527993 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.537575 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7b86f4c74d-dlwfm"] Oct 01 09:03:18 crc kubenswrapper[4837]: I1001 09:03:18.816811 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:03:18 crc kubenswrapper[4837]: E1001 09:03:18.817309 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:19 crc kubenswrapper[4837]: I1001 09:03:19.835799 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" path="/var/lib/kubelet/pods/719662c0-7881-465a-b1ab-dffb9b07f386/volumes" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.716242 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zdpwn"] Oct 01 09:03:28 crc kubenswrapper[4837]: E1001 09:03:28.716985 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="dnsmasq-dns" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.716999 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="dnsmasq-dns" Oct 01 09:03:28 crc kubenswrapper[4837]: E1001 09:03:28.717018 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-httpd" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717024 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-httpd" Oct 01 09:03:28 crc kubenswrapper[4837]: E1001 09:03:28.717039 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="init" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717044 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="init" Oct 01 09:03:28 crc kubenswrapper[4837]: E1001 09:03:28.717066 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-api" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717072 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-api" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717222 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f5b76b2-a3e5-4f6d-a5c0-b2933a46c84c" containerName="dnsmasq-dns" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717234 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-api" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717246 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="719662c0-7881-465a-b1ab-dffb9b07f386" containerName="neutron-httpd" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.717827 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.721981 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.722231 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.722280 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6jq5k" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.722323 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.723078 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.747461 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zdpwn"] Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824483 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824535 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824590 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824617 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cmjq\" (UniqueName: \"kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824707 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.824991 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.825732 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.827284 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.838475 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.926758 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfvd\" (UniqueName: \"kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.926814 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cmjq\" (UniqueName: \"kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.926833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.926856 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.926949 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927001 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927062 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927091 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927175 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927201 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.927225 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.929898 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.934826 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.935999 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.936928 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.937734 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.950155 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:28 crc kubenswrapper[4837]: I1001 09:03:28.950777 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cmjq\" (UniqueName: \"kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq\") pod \"swift-ring-rebalance-zdpwn\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.028668 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.028734 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.028783 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.028832 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfvd\" (UniqueName: \"kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.028854 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.029787 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.029799 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.029810 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.030247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.039408 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.043888 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfvd\" (UniqueName: \"kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd\") pod \"dnsmasq-dns-6669cbdcf7-pxd6c\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.148499 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.508799 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zdpwn"] Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.589152 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zdpwn" event={"ID":"e55925e1-26b2-42da-a0b6-dacb251e66ed","Type":"ContainerStarted","Data":"a0cf4d941e4e566531f218eb4528ddcf9ad6d25fa458cba0145d26c410d3144a"} Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.680860 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:03:29 crc kubenswrapper[4837]: W1001 09:03:29.689810 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf99ca47f_4a95_4a1f_8342_ff5e2d6af2b0.slice/crio-e6db9d86e7fca29c3af63feb82b2dd8d6b36112fbbf002864a8afc84441a2177 WatchSource:0}: Error finding container e6db9d86e7fca29c3af63feb82b2dd8d6b36112fbbf002864a8afc84441a2177: Status 404 returned error can't find the container with id e6db9d86e7fca29c3af63feb82b2dd8d6b36112fbbf002864a8afc84441a2177 Oct 01 09:03:29 crc kubenswrapper[4837]: I1001 09:03:29.819049 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:03:29 crc kubenswrapper[4837]: E1001 09:03:29.819329 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.600551 4837 generic.go:334] "Generic (PLEG): container finished" podID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerID="8a5b24f86fc0d377779188d082a98168656b387ccd435b2ebc5fa92745516fe1" exitCode=0 Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.600892 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" event={"ID":"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0","Type":"ContainerDied","Data":"8a5b24f86fc0d377779188d082a98168656b387ccd435b2ebc5fa92745516fe1"} Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.600923 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" event={"ID":"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0","Type":"ContainerStarted","Data":"e6db9d86e7fca29c3af63feb82b2dd8d6b36112fbbf002864a8afc84441a2177"} Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.872902 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.877794 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.881020 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 01 09:03:30 crc kubenswrapper[4837]: I1001 09:03:30.883156 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.081807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.081893 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvdfj\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.081933 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.081949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.081973 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.082075 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183633 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183745 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvdfj\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183784 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183810 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183838 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.183911 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.184237 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.184268 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.190125 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.190791 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.191528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.217457 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvdfj\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj\") pod \"swift-proxy-586d7cf9c9-vh75l\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.507608 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.613570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" event={"ID":"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0","Type":"ContainerStarted","Data":"723bbf703ed1d28b0c5040eb3899764766b5b06150afd86006d625d0e08ccd2c"} Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.613755 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:31 crc kubenswrapper[4837]: I1001 09:03:31.633843 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" podStartSLOduration=3.633820101 podStartE2EDuration="3.633820101s" podCreationTimestamp="2025-10-01 09:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:03:31.629096434 +0000 UTC m=+7068.470703889" watchObservedRunningTime="2025-10-01 09:03:31.633820101 +0000 UTC m=+7068.475427566" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.745190 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-b44b6c8b7-fh9pg"] Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.746512 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.748710 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.753776 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.767127 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b44b6c8b7-fh9pg"] Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914620 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-combined-ca-bundle\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914679 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-public-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914767 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqw8l\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-kube-api-access-vqw8l\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914816 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-etc-swift\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914907 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-log-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914929 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-run-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914953 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-internal-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:32 crc kubenswrapper[4837]: I1001 09:03:32.914974 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-config-data\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017415 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-log-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017454 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-run-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017468 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-internal-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017494 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-config-data\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-combined-ca-bundle\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017556 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-public-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017598 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqw8l\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-kube-api-access-vqw8l\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.017645 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-etc-swift\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.018853 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-log-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.018856 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2b4b56f8-7b62-4209-9683-8a6c842eabe5-run-httpd\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.023559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-etc-swift\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.023582 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-internal-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.024489 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-public-tls-certs\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.028263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-config-data\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.035850 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4b56f8-7b62-4209-9683-8a6c842eabe5-combined-ca-bundle\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.035975 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqw8l\" (UniqueName: \"kubernetes.io/projected/2b4b56f8-7b62-4209-9683-8a6c842eabe5-kube-api-access-vqw8l\") pod \"swift-proxy-b44b6c8b7-fh9pg\" (UID: \"2b4b56f8-7b62-4209-9683-8a6c842eabe5\") " pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.073440 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:33 crc kubenswrapper[4837]: W1001 09:03:33.896219 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b4b56f8_7b62_4209_9683_8a6c842eabe5.slice/crio-69274ec2ac2cabd4c05eb5bd110e0792b0b1111c01f8a3f2d97bc2f023b33aff WatchSource:0}: Error finding container 69274ec2ac2cabd4c05eb5bd110e0792b0b1111c01f8a3f2d97bc2f023b33aff: Status 404 returned error can't find the container with id 69274ec2ac2cabd4c05eb5bd110e0792b0b1111c01f8a3f2d97bc2f023b33aff Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.920945 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b44b6c8b7-fh9pg"] Oct 01 09:03:33 crc kubenswrapper[4837]: I1001 09:03:33.990477 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:34 crc kubenswrapper[4837]: W1001 09:03:34.004197 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfae00536_707d_4a46_9b97_cc6d072b4c84.slice/crio-887b49633e75c337c10a8e707abc7a60503827e9040c018f9ed4ae8082556bfa WatchSource:0}: Error finding container 887b49633e75c337c10a8e707abc7a60503827e9040c018f9ed4ae8082556bfa: Status 404 returned error can't find the container with id 887b49633e75c337c10a8e707abc7a60503827e9040c018f9ed4ae8082556bfa Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.647973 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zdpwn" event={"ID":"e55925e1-26b2-42da-a0b6-dacb251e66ed","Type":"ContainerStarted","Data":"4b9a7ff4a5e29f33a867427289d69fb34ef1aca2e3a5c9a1a4c1cab4d46669b1"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.672279 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zdpwn" podStartSLOduration=2.80902136 podStartE2EDuration="6.67226138s" podCreationTimestamp="2025-10-01 09:03:28 +0000 UTC" firstStartedPulling="2025-10-01 09:03:29.515215558 +0000 UTC m=+7066.356823013" lastFinishedPulling="2025-10-01 09:03:33.378455578 +0000 UTC m=+7070.220063033" observedRunningTime="2025-10-01 09:03:34.66907079 +0000 UTC m=+7071.510678245" watchObservedRunningTime="2025-10-01 09:03:34.67226138 +0000 UTC m=+7071.513868835" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.677089 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerStarted","Data":"bf0360a5d905e655ad635244886ecaf3e0f2274876e37cf5b2f623b1a811f1dc"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.677156 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerStarted","Data":"e7a031931eb2ffcef32bf616b9af7992f76d0ebf4f00bb556a2d3893f782f180"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.677172 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerStarted","Data":"887b49633e75c337c10a8e707abc7a60503827e9040c018f9ed4ae8082556bfa"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.678315 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.678354 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.683444 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" event={"ID":"2b4b56f8-7b62-4209-9683-8a6c842eabe5","Type":"ContainerStarted","Data":"967aa942ae3633d4a0feb04337e41d153789e425d7d76c27e0fec4e4fda28caa"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.683513 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" event={"ID":"2b4b56f8-7b62-4209-9683-8a6c842eabe5","Type":"ContainerStarted","Data":"85aec65c27dd8c9c82a7bb20a6b4dff303dba5b8f2496857d4185d2fda06ae51"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.683524 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" event={"ID":"2b4b56f8-7b62-4209-9683-8a6c842eabe5","Type":"ContainerStarted","Data":"69274ec2ac2cabd4c05eb5bd110e0792b0b1111c01f8a3f2d97bc2f023b33aff"} Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.684552 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.684574 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.710081 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-586d7cf9c9-vh75l" podStartSLOduration=4.710058088 podStartE2EDuration="4.710058088s" podCreationTimestamp="2025-10-01 09:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:03:34.70406916 +0000 UTC m=+7071.545676615" watchObservedRunningTime="2025-10-01 09:03:34.710058088 +0000 UTC m=+7071.551665543" Oct 01 09:03:34 crc kubenswrapper[4837]: I1001 09:03:34.734169 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" podStartSLOduration=2.734144399 podStartE2EDuration="2.734144399s" podCreationTimestamp="2025-10-01 09:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:03:34.721782446 +0000 UTC m=+7071.563389901" watchObservedRunningTime="2025-10-01 09:03:34.734144399 +0000 UTC m=+7071.575751864" Oct 01 09:03:38 crc kubenswrapper[4837]: I1001 09:03:38.729645 4837 generic.go:334] "Generic (PLEG): container finished" podID="e55925e1-26b2-42da-a0b6-dacb251e66ed" containerID="4b9a7ff4a5e29f33a867427289d69fb34ef1aca2e3a5c9a1a4c1cab4d46669b1" exitCode=0 Oct 01 09:03:38 crc kubenswrapper[4837]: I1001 09:03:38.729739 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zdpwn" event={"ID":"e55925e1-26b2-42da-a0b6-dacb251e66ed","Type":"ContainerDied","Data":"4b9a7ff4a5e29f33a867427289d69fb34ef1aca2e3a5c9a1a4c1cab4d46669b1"} Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.151013 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.241007 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.241321 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="dnsmasq-dns" containerID="cri-o://26289cb6f00e22a8b002ea5b7e6fb3e1b300f8d23fa798f2de0896520262a3a7" gracePeriod=10 Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.744017 4837 generic.go:334] "Generic (PLEG): container finished" podID="3b19b130-2f32-40fe-afac-80cae13880f7" containerID="26289cb6f00e22a8b002ea5b7e6fb3e1b300f8d23fa798f2de0896520262a3a7" exitCode=0 Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.744108 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" event={"ID":"3b19b130-2f32-40fe-afac-80cae13880f7","Type":"ContainerDied","Data":"26289cb6f00e22a8b002ea5b7e6fb3e1b300f8d23fa798f2de0896520262a3a7"} Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.744445 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" event={"ID":"3b19b130-2f32-40fe-afac-80cae13880f7","Type":"ContainerDied","Data":"17f79d6ead13974668bb90313a10024fe27cecda7ff029bc00f36b2f13493671"} Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.744461 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f79d6ead13974668bb90313a10024fe27cecda7ff029bc00f36b2f13493671" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.781829 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.850265 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb\") pod \"3b19b130-2f32-40fe-afac-80cae13880f7\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.850359 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc\") pod \"3b19b130-2f32-40fe-afac-80cae13880f7\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.850530 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xmph\" (UniqueName: \"kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph\") pod \"3b19b130-2f32-40fe-afac-80cae13880f7\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.850558 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config\") pod \"3b19b130-2f32-40fe-afac-80cae13880f7\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.850575 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb\") pod \"3b19b130-2f32-40fe-afac-80cae13880f7\" (UID: \"3b19b130-2f32-40fe-afac-80cae13880f7\") " Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.860250 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph" (OuterVolumeSpecName: "kube-api-access-5xmph") pod "3b19b130-2f32-40fe-afac-80cae13880f7" (UID: "3b19b130-2f32-40fe-afac-80cae13880f7"). InnerVolumeSpecName "kube-api-access-5xmph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.903440 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config" (OuterVolumeSpecName: "config") pod "3b19b130-2f32-40fe-afac-80cae13880f7" (UID: "3b19b130-2f32-40fe-afac-80cae13880f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.906604 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3b19b130-2f32-40fe-afac-80cae13880f7" (UID: "3b19b130-2f32-40fe-afac-80cae13880f7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.915754 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3b19b130-2f32-40fe-afac-80cae13880f7" (UID: "3b19b130-2f32-40fe-afac-80cae13880f7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.940107 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b19b130-2f32-40fe-afac-80cae13880f7" (UID: "3b19b130-2f32-40fe-afac-80cae13880f7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.956433 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.956503 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.956520 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xmph\" (UniqueName: \"kubernetes.io/projected/3b19b130-2f32-40fe-afac-80cae13880f7-kube-api-access-5xmph\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.956568 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:39 crc kubenswrapper[4837]: I1001 09:03:39.956581 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b19b130-2f32-40fe-afac-80cae13880f7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.060316 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159449 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159575 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159733 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159773 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cmjq\" (UniqueName: \"kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159897 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.159954 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift\") pod \"e55925e1-26b2-42da-a0b6-dacb251e66ed\" (UID: \"e55925e1-26b2-42da-a0b6-dacb251e66ed\") " Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.160328 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.160446 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.161599 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.165876 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq" (OuterVolumeSpecName: "kube-api-access-5cmjq") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "kube-api-access-5cmjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.167072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.186850 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.198508 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts" (OuterVolumeSpecName: "scripts") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.206083 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e55925e1-26b2-42da-a0b6-dacb251e66ed" (UID: "e55925e1-26b2-42da-a0b6-dacb251e66ed"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.262968 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.263014 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.263031 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cmjq\" (UniqueName: \"kubernetes.io/projected/e55925e1-26b2-42da-a0b6-dacb251e66ed-kube-api-access-5cmjq\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.263045 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e55925e1-26b2-42da-a0b6-dacb251e66ed-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.263056 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e55925e1-26b2-42da-a0b6-dacb251e66ed-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.263066 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e55925e1-26b2-42da-a0b6-dacb251e66ed-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.757799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zdpwn" event={"ID":"e55925e1-26b2-42da-a0b6-dacb251e66ed","Type":"ContainerDied","Data":"a0cf4d941e4e566531f218eb4528ddcf9ad6d25fa458cba0145d26c410d3144a"} Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.757829 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.757851 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0cf4d941e4e566531f218eb4528ddcf9ad6d25fa458cba0145d26c410d3144a" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.757952 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zdpwn" Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.825366 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:03:40 crc kubenswrapper[4837]: I1001 09:03:40.835274 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869fcf5fcf-kwxkg"] Oct 01 09:03:41 crc kubenswrapper[4837]: I1001 09:03:41.511195 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:41 crc kubenswrapper[4837]: I1001 09:03:41.511298 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:41 crc kubenswrapper[4837]: I1001 09:03:41.829209 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" path="/var/lib/kubelet/pods/3b19b130-2f32-40fe-afac-80cae13880f7/volumes" Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.081109 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.082773 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b44b6c8b7-fh9pg" Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.189934 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.190434 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-586d7cf9c9-vh75l" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-server" containerID="cri-o://bf0360a5d905e655ad635244886ecaf3e0f2274876e37cf5b2f623b1a811f1dc" gracePeriod=30 Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.190434 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-586d7cf9c9-vh75l" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-httpd" containerID="cri-o://e7a031931eb2ffcef32bf616b9af7992f76d0ebf4f00bb556a2d3893f782f180" gracePeriod=30 Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.802449 4837 generic.go:334] "Generic (PLEG): container finished" podID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerID="bf0360a5d905e655ad635244886ecaf3e0f2274876e37cf5b2f623b1a811f1dc" exitCode=0 Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.802842 4837 generic.go:334] "Generic (PLEG): container finished" podID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerID="e7a031931eb2ffcef32bf616b9af7992f76d0ebf4f00bb556a2d3893f782f180" exitCode=0 Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.802618 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerDied","Data":"bf0360a5d905e655ad635244886ecaf3e0f2274876e37cf5b2f623b1a811f1dc"} Oct 01 09:03:43 crc kubenswrapper[4837]: I1001 09:03:43.803202 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerDied","Data":"e7a031931eb2ffcef32bf616b9af7992f76d0ebf4f00bb556a2d3893f782f180"} Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.303614 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.447860 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448234 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvdfj\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448260 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448322 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448529 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448561 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle\") pod \"fae00536-707d-4a46-9b97-cc6d072b4c84\" (UID: \"fae00536-707d-4a46-9b97-cc6d072b4c84\") " Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.448966 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.449137 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.449371 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.449386 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fae00536-707d-4a46-9b97-cc6d072b4c84-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.454085 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj" (OuterVolumeSpecName: "kube-api-access-hvdfj") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "kube-api-access-hvdfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.456869 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.510195 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.523474 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data" (OuterVolumeSpecName: "config-data") pod "fae00536-707d-4a46-9b97-cc6d072b4c84" (UID: "fae00536-707d-4a46-9b97-cc6d072b4c84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.550596 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.550639 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae00536-707d-4a46-9b97-cc6d072b4c84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.550651 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvdfj\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-kube-api-access-hvdfj\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.550662 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fae00536-707d-4a46-9b97-cc6d072b4c84-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.610747 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-869fcf5fcf-kwxkg" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.50:5353: i/o timeout" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.815576 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:03:44 crc kubenswrapper[4837]: E1001 09:03:44.816005 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.817241 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-586d7cf9c9-vh75l" event={"ID":"fae00536-707d-4a46-9b97-cc6d072b4c84","Type":"ContainerDied","Data":"887b49633e75c337c10a8e707abc7a60503827e9040c018f9ed4ae8082556bfa"} Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.817314 4837 scope.go:117] "RemoveContainer" containerID="bf0360a5d905e655ad635244886ecaf3e0f2274876e37cf5b2f623b1a811f1dc" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.817319 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-586d7cf9c9-vh75l" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.844349 4837 scope.go:117] "RemoveContainer" containerID="e7a031931eb2ffcef32bf616b9af7992f76d0ebf4f00bb556a2d3893f782f180" Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.872591 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:44 crc kubenswrapper[4837]: I1001 09:03:44.880069 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-586d7cf9c9-vh75l"] Oct 01 09:03:45 crc kubenswrapper[4837]: I1001 09:03:45.830567 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" path="/var/lib/kubelet/pods/fae00536-707d-4a46-9b97-cc6d072b4c84/volumes" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.090199 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-96d22"] Oct 01 09:03:49 crc kubenswrapper[4837]: E1001 09:03:49.090991 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="init" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091013 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="init" Oct 01 09:03:49 crc kubenswrapper[4837]: E1001 09:03:49.091053 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="dnsmasq-dns" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091063 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="dnsmasq-dns" Oct 01 09:03:49 crc kubenswrapper[4837]: E1001 09:03:49.091076 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-httpd" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091084 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-httpd" Oct 01 09:03:49 crc kubenswrapper[4837]: E1001 09:03:49.091097 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55925e1-26b2-42da-a0b6-dacb251e66ed" containerName="swift-ring-rebalance" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091105 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55925e1-26b2-42da-a0b6-dacb251e66ed" containerName="swift-ring-rebalance" Oct 01 09:03:49 crc kubenswrapper[4837]: E1001 09:03:49.091130 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-server" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091137 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-server" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091358 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-server" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091374 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b19b130-2f32-40fe-afac-80cae13880f7" containerName="dnsmasq-dns" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091393 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae00536-707d-4a46-9b97-cc6d072b4c84" containerName="proxy-httpd" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.091433 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e55925e1-26b2-42da-a0b6-dacb251e66ed" containerName="swift-ring-rebalance" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.092101 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-96d22" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.103068 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-96d22"] Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.144952 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2p2r\" (UniqueName: \"kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r\") pod \"cinder-db-create-96d22\" (UID: \"007a2a06-3b17-43b4-b070-df7291499232\") " pod="openstack/cinder-db-create-96d22" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.246771 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2p2r\" (UniqueName: \"kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r\") pod \"cinder-db-create-96d22\" (UID: \"007a2a06-3b17-43b4-b070-df7291499232\") " pod="openstack/cinder-db-create-96d22" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.269931 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2p2r\" (UniqueName: \"kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r\") pod \"cinder-db-create-96d22\" (UID: \"007a2a06-3b17-43b4-b070-df7291499232\") " pod="openstack/cinder-db-create-96d22" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.420841 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-96d22" Oct 01 09:03:49 crc kubenswrapper[4837]: I1001 09:03:49.914447 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-96d22"] Oct 01 09:03:49 crc kubenswrapper[4837]: W1001 09:03:49.920907 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod007a2a06_3b17_43b4_b070_df7291499232.slice/crio-e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9 WatchSource:0}: Error finding container e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9: Status 404 returned error can't find the container with id e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9 Oct 01 09:03:50 crc kubenswrapper[4837]: I1001 09:03:50.883910 4837 generic.go:334] "Generic (PLEG): container finished" podID="007a2a06-3b17-43b4-b070-df7291499232" containerID="1d49f5cc9fdaf44790f2e29467bfe6ef969b66744323a2e05955d0ea02578963" exitCode=0 Oct 01 09:03:50 crc kubenswrapper[4837]: I1001 09:03:50.884078 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-96d22" event={"ID":"007a2a06-3b17-43b4-b070-df7291499232","Type":"ContainerDied","Data":"1d49f5cc9fdaf44790f2e29467bfe6ef969b66744323a2e05955d0ea02578963"} Oct 01 09:03:50 crc kubenswrapper[4837]: I1001 09:03:50.884494 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-96d22" event={"ID":"007a2a06-3b17-43b4-b070-df7291499232","Type":"ContainerStarted","Data":"e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9"} Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.304643 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-96d22" Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.408908 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2p2r\" (UniqueName: \"kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r\") pod \"007a2a06-3b17-43b4-b070-df7291499232\" (UID: \"007a2a06-3b17-43b4-b070-df7291499232\") " Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.417081 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r" (OuterVolumeSpecName: "kube-api-access-p2p2r") pod "007a2a06-3b17-43b4-b070-df7291499232" (UID: "007a2a06-3b17-43b4-b070-df7291499232"). InnerVolumeSpecName "kube-api-access-p2p2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.511452 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2p2r\" (UniqueName: \"kubernetes.io/projected/007a2a06-3b17-43b4-b070-df7291499232-kube-api-access-p2p2r\") on node \"crc\" DevicePath \"\"" Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.910899 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-96d22" event={"ID":"007a2a06-3b17-43b4-b070-df7291499232","Type":"ContainerDied","Data":"e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9"} Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.910958 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9ed01e586cb9c9f40b71ed9de55c8647c44ac614156da26f3ecb6b3ca3e75f9" Oct 01 09:03:52 crc kubenswrapper[4837]: I1001 09:03:52.911028 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-96d22" Oct 01 09:03:57 crc kubenswrapper[4837]: I1001 09:03:57.816656 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:03:57 crc kubenswrapper[4837]: E1001 09:03:57.817519 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.193479 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-31a1-account-create-sfpml"] Oct 01 09:03:59 crc kubenswrapper[4837]: E1001 09:03:59.194718 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="007a2a06-3b17-43b4-b070-df7291499232" containerName="mariadb-database-create" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.194740 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="007a2a06-3b17-43b4-b070-df7291499232" containerName="mariadb-database-create" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.194972 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="007a2a06-3b17-43b4-b070-df7291499232" containerName="mariadb-database-create" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.195782 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.199813 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.203800 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-31a1-account-create-sfpml"] Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.260792 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb5tm\" (UniqueName: \"kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm\") pod \"cinder-31a1-account-create-sfpml\" (UID: \"011cca03-8337-4be1-9b4d-c2d3862a0db5\") " pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.363262 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb5tm\" (UniqueName: \"kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm\") pod \"cinder-31a1-account-create-sfpml\" (UID: \"011cca03-8337-4be1-9b4d-c2d3862a0db5\") " pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.385867 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb5tm\" (UniqueName: \"kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm\") pod \"cinder-31a1-account-create-sfpml\" (UID: \"011cca03-8337-4be1-9b4d-c2d3862a0db5\") " pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.516016 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:03:59 crc kubenswrapper[4837]: I1001 09:03:59.996407 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-31a1-account-create-sfpml"] Oct 01 09:04:00 crc kubenswrapper[4837]: W1001 09:04:00.000142 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod011cca03_8337_4be1_9b4d_c2d3862a0db5.slice/crio-d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219 WatchSource:0}: Error finding container d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219: Status 404 returned error can't find the container with id d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219 Oct 01 09:04:01 crc kubenswrapper[4837]: I1001 09:04:01.006405 4837 generic.go:334] "Generic (PLEG): container finished" podID="011cca03-8337-4be1-9b4d-c2d3862a0db5" containerID="b44dd1c021485d8bdfb1f9f176ab5210c897de7b373ac5af4b8548cc51edf723" exitCode=0 Oct 01 09:04:01 crc kubenswrapper[4837]: I1001 09:04:01.006529 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31a1-account-create-sfpml" event={"ID":"011cca03-8337-4be1-9b4d-c2d3862a0db5","Type":"ContainerDied","Data":"b44dd1c021485d8bdfb1f9f176ab5210c897de7b373ac5af4b8548cc51edf723"} Oct 01 09:04:01 crc kubenswrapper[4837]: I1001 09:04:01.006840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31a1-account-create-sfpml" event={"ID":"011cca03-8337-4be1-9b4d-c2d3862a0db5","Type":"ContainerStarted","Data":"d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219"} Oct 01 09:04:02 crc kubenswrapper[4837]: I1001 09:04:02.447902 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:04:02 crc kubenswrapper[4837]: I1001 09:04:02.527083 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb5tm\" (UniqueName: \"kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm\") pod \"011cca03-8337-4be1-9b4d-c2d3862a0db5\" (UID: \"011cca03-8337-4be1-9b4d-c2d3862a0db5\") " Oct 01 09:04:02 crc kubenswrapper[4837]: I1001 09:04:02.533536 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm" (OuterVolumeSpecName: "kube-api-access-tb5tm") pod "011cca03-8337-4be1-9b4d-c2d3862a0db5" (UID: "011cca03-8337-4be1-9b4d-c2d3862a0db5"). InnerVolumeSpecName "kube-api-access-tb5tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:04:02 crc kubenswrapper[4837]: I1001 09:04:02.630435 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb5tm\" (UniqueName: \"kubernetes.io/projected/011cca03-8337-4be1-9b4d-c2d3862a0db5-kube-api-access-tb5tm\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:03 crc kubenswrapper[4837]: I1001 09:04:03.029985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-31a1-account-create-sfpml" event={"ID":"011cca03-8337-4be1-9b4d-c2d3862a0db5","Type":"ContainerDied","Data":"d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219"} Oct 01 09:04:03 crc kubenswrapper[4837]: I1001 09:04:03.030345 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8c46f90c5776769ee47e10006d6cd1445522181ee661be5c3d7d7be9fbdb219" Oct 01 09:04:03 crc kubenswrapper[4837]: I1001 09:04:03.030112 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-31a1-account-create-sfpml" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.360256 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vzspd"] Oct 01 09:04:04 crc kubenswrapper[4837]: E1001 09:04:04.362280 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011cca03-8337-4be1-9b4d-c2d3862a0db5" containerName="mariadb-account-create" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.362418 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="011cca03-8337-4be1-9b4d-c2d3862a0db5" containerName="mariadb-account-create" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.362822 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="011cca03-8337-4be1-9b4d-c2d3862a0db5" containerName="mariadb-account-create" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.363940 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.367204 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.367514 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v2k57" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.367908 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vzspd"] Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.368067 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476122 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476201 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476359 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476388 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476447 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.476473 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnqw4\" (UniqueName: \"kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581218 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581298 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnqw4\" (UniqueName: \"kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581351 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581415 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581540 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.581592 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.588434 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.590309 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.590447 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.593017 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.604133 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnqw4\" (UniqueName: \"kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4\") pod \"cinder-db-sync-vzspd\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:04 crc kubenswrapper[4837]: I1001 09:04:04.701659 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:05 crc kubenswrapper[4837]: I1001 09:04:05.148860 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vzspd"] Oct 01 09:04:06 crc kubenswrapper[4837]: I1001 09:04:06.056935 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vzspd" event={"ID":"d5a7970a-45f3-48f2-adec-2a19833c2fe5","Type":"ContainerStarted","Data":"31abc7d66afd7efa8a8f294420111e1538b1e6815a0c242354cc34ff3ea4ac13"} Oct 01 09:04:06 crc kubenswrapper[4837]: I1001 09:04:06.751649 4837 scope.go:117] "RemoveContainer" containerID="d710015d36f33963b29c241e3ce92ee592a58e400662174ab274488ac0229639" Oct 01 09:04:06 crc kubenswrapper[4837]: I1001 09:04:06.780208 4837 scope.go:117] "RemoveContainer" containerID="2e31be9975305e39a5602ea5a2b1bfee78f1113567b1ccb6803ce3073163ed37" Oct 01 09:04:11 crc kubenswrapper[4837]: I1001 09:04:11.818459 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:04:11 crc kubenswrapper[4837]: E1001 09:04:11.819560 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:04:26 crc kubenswrapper[4837]: I1001 09:04:26.241082 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vzspd" event={"ID":"d5a7970a-45f3-48f2-adec-2a19833c2fe5","Type":"ContainerStarted","Data":"1a707fc0da32a6a48304a5601857ee0a448895c615ec0b15a4454cfdfd7e442c"} Oct 01 09:04:26 crc kubenswrapper[4837]: I1001 09:04:26.269877 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vzspd" podStartSLOduration=2.755794704 podStartE2EDuration="22.269855724s" podCreationTimestamp="2025-10-01 09:04:04 +0000 UTC" firstStartedPulling="2025-10-01 09:04:05.179450952 +0000 UTC m=+7102.021058417" lastFinishedPulling="2025-10-01 09:04:24.693511972 +0000 UTC m=+7121.535119437" observedRunningTime="2025-10-01 09:04:26.261722645 +0000 UTC m=+7123.103330100" watchObservedRunningTime="2025-10-01 09:04:26.269855724 +0000 UTC m=+7123.111463179" Oct 01 09:04:26 crc kubenswrapper[4837]: I1001 09:04:26.816539 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:04:26 crc kubenswrapper[4837]: E1001 09:04:26.816927 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:04:28 crc kubenswrapper[4837]: I1001 09:04:28.266760 4837 generic.go:334] "Generic (PLEG): container finished" podID="d5a7970a-45f3-48f2-adec-2a19833c2fe5" containerID="1a707fc0da32a6a48304a5601857ee0a448895c615ec0b15a4454cfdfd7e442c" exitCode=0 Oct 01 09:04:28 crc kubenswrapper[4837]: I1001 09:04:28.266898 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vzspd" event={"ID":"d5a7970a-45f3-48f2-adec-2a19833c2fe5","Type":"ContainerDied","Data":"1a707fc0da32a6a48304a5601857ee0a448895c615ec0b15a4454cfdfd7e442c"} Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.601943 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682763 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682866 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnqw4\" (UniqueName: \"kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682914 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682934 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682950 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.682983 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data\") pod \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\" (UID: \"d5a7970a-45f3-48f2-adec-2a19833c2fe5\") " Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.683649 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.684419 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d5a7970a-45f3-48f2-adec-2a19833c2fe5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.688931 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.688963 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts" (OuterVolumeSpecName: "scripts") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.689033 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4" (OuterVolumeSpecName: "kube-api-access-cnqw4") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "kube-api-access-cnqw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.706867 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.732572 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data" (OuterVolumeSpecName: "config-data") pod "d5a7970a-45f3-48f2-adec-2a19833c2fe5" (UID: "d5a7970a-45f3-48f2-adec-2a19833c2fe5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.786123 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.786168 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.786183 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.786199 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5a7970a-45f3-48f2-adec-2a19833c2fe5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:29 crc kubenswrapper[4837]: I1001 09:04:29.786212 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnqw4\" (UniqueName: \"kubernetes.io/projected/d5a7970a-45f3-48f2-adec-2a19833c2fe5-kube-api-access-cnqw4\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.293664 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vzspd" event={"ID":"d5a7970a-45f3-48f2-adec-2a19833c2fe5","Type":"ContainerDied","Data":"31abc7d66afd7efa8a8f294420111e1538b1e6815a0c242354cc34ff3ea4ac13"} Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.293784 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31abc7d66afd7efa8a8f294420111e1538b1e6815a0c242354cc34ff3ea4ac13" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.293813 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vzspd" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.680965 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:04:30 crc kubenswrapper[4837]: E1001 09:04:30.681584 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a7970a-45f3-48f2-adec-2a19833c2fe5" containerName="cinder-db-sync" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.681602 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a7970a-45f3-48f2-adec-2a19833c2fe5" containerName="cinder-db-sync" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.681848 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5a7970a-45f3-48f2-adec-2a19833c2fe5" containerName="cinder-db-sync" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.683875 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.710746 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.807391 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5m4j\" (UniqueName: \"kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.807565 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.807616 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.807711 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.807758 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.897714 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.900206 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.908277 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.908335 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v2k57" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.908274 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.909784 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.909868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.909904 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.909931 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.910025 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5m4j\" (UniqueName: \"kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.911770 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.911841 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.912393 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.912573 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.912968 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.926252 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:30 crc kubenswrapper[4837]: I1001 09:04:30.947523 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5m4j\" (UniqueName: \"kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j\") pod \"dnsmasq-dns-66b464cb87-qcrc7\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012177 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012276 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012316 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj742\" (UniqueName: \"kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012344 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012391 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012431 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.012525 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.020289 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117044 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117105 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117139 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj742\" (UniqueName: \"kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117160 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117184 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117210 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.117252 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.118636 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.119136 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.123476 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.124818 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.126304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.130430 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.140398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj742\" (UniqueName: \"kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742\") pod \"cinder-api-0\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.287651 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.394137 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:04:31 crc kubenswrapper[4837]: I1001 09:04:31.837369 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:31 crc kubenswrapper[4837]: W1001 09:04:31.844048 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e43ee6b_ea30_446f_9c87_8bc8b319423e.slice/crio-84aaeb743ce2d5247f68b1c4a82186b2830cf76718ecb8a5081317275f297405 WatchSource:0}: Error finding container 84aaeb743ce2d5247f68b1c4a82186b2830cf76718ecb8a5081317275f297405: Status 404 returned error can't find the container with id 84aaeb743ce2d5247f68b1c4a82186b2830cf76718ecb8a5081317275f297405 Oct 01 09:04:32 crc kubenswrapper[4837]: I1001 09:04:32.345262 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerStarted","Data":"84aaeb743ce2d5247f68b1c4a82186b2830cf76718ecb8a5081317275f297405"} Oct 01 09:04:32 crc kubenswrapper[4837]: I1001 09:04:32.348607 4837 generic.go:334] "Generic (PLEG): container finished" podID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerID="f842a1e3b920a2313fe5810dfec9ba48b40514c8c601dfe1edfe5c246785e96d" exitCode=0 Oct 01 09:04:32 crc kubenswrapper[4837]: I1001 09:04:32.348679 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" event={"ID":"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d","Type":"ContainerDied","Data":"f842a1e3b920a2313fe5810dfec9ba48b40514c8c601dfe1edfe5c246785e96d"} Oct 01 09:04:32 crc kubenswrapper[4837]: I1001 09:04:32.348736 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" event={"ID":"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d","Type":"ContainerStarted","Data":"e9994ddc852466630c5ad58e2ec1f88642f8f73322db1adc50e101917959703a"} Oct 01 09:04:32 crc kubenswrapper[4837]: I1001 09:04:32.723569 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:33 crc kubenswrapper[4837]: I1001 09:04:33.362028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" event={"ID":"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d","Type":"ContainerStarted","Data":"8678c653aa5843282f890228a0a4127405fb6217649ad0916d4d7720c4db631d"} Oct 01 09:04:33 crc kubenswrapper[4837]: I1001 09:04:33.362380 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:33 crc kubenswrapper[4837]: I1001 09:04:33.368803 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerStarted","Data":"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53"} Oct 01 09:04:33 crc kubenswrapper[4837]: I1001 09:04:33.848004 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" podStartSLOduration=3.84798525 podStartE2EDuration="3.84798525s" podCreationTimestamp="2025-10-01 09:04:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:04:33.390601814 +0000 UTC m=+7130.232209269" watchObservedRunningTime="2025-10-01 09:04:33.84798525 +0000 UTC m=+7130.689592715" Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.380068 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api-log" containerID="cri-o://d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" gracePeriod=30 Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.380464 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerStarted","Data":"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1"} Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.380538 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.380581 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api" containerID="cri-o://316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" gracePeriod=30 Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.413737 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.413716687 podStartE2EDuration="4.413716687s" podCreationTimestamp="2025-10-01 09:04:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:04:34.407822682 +0000 UTC m=+7131.249430137" watchObservedRunningTime="2025-10-01 09:04:34.413716687 +0000 UTC m=+7131.255324142" Oct 01 09:04:34 crc kubenswrapper[4837]: I1001 09:04:34.991953 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035143 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035207 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035283 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035395 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035522 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035673 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj742\" (UniqueName: \"kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035753 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id\") pod \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\" (UID: \"7e43ee6b-ea30-446f-9c87-8bc8b319423e\") " Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.035958 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs" (OuterVolumeSpecName: "logs") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.036426 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e43ee6b-ea30-446f-9c87-8bc8b319423e-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.036483 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.042126 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.042863 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts" (OuterVolumeSpecName: "scripts") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.064053 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742" (OuterVolumeSpecName: "kube-api-access-sj742") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "kube-api-access-sj742". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.067661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.101504 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data" (OuterVolumeSpecName: "config-data") pod "7e43ee6b-ea30-446f-9c87-8bc8b319423e" (UID: "7e43ee6b-ea30-446f-9c87-8bc8b319423e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138407 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138460 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj742\" (UniqueName: \"kubernetes.io/projected/7e43ee6b-ea30-446f-9c87-8bc8b319423e-kube-api-access-sj742\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138476 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7e43ee6b-ea30-446f-9c87-8bc8b319423e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138490 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138502 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.138515 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e43ee6b-ea30-446f-9c87-8bc8b319423e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.388422 4837 generic.go:334] "Generic (PLEG): container finished" podID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerID="316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" exitCode=0 Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.388866 4837 generic.go:334] "Generic (PLEG): container finished" podID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerID="d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" exitCode=143 Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.388514 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.388481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerDied","Data":"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1"} Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.389009 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerDied","Data":"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53"} Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.389038 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7e43ee6b-ea30-446f-9c87-8bc8b319423e","Type":"ContainerDied","Data":"84aaeb743ce2d5247f68b1c4a82186b2830cf76718ecb8a5081317275f297405"} Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.389064 4837 scope.go:117] "RemoveContainer" containerID="316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.407799 4837 scope.go:117] "RemoveContainer" containerID="d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.420712 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.429349 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.454961 4837 scope.go:117] "RemoveContainer" containerID="316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" Oct 01 09:04:35 crc kubenswrapper[4837]: E1001 09:04:35.455662 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1\": container with ID starting with 316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1 not found: ID does not exist" containerID="316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.455822 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1"} err="failed to get container status \"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1\": rpc error: code = NotFound desc = could not find container \"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1\": container with ID starting with 316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1 not found: ID does not exist" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.455986 4837 scope.go:117] "RemoveContainer" containerID="d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" Oct 01 09:04:35 crc kubenswrapper[4837]: E1001 09:04:35.456431 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53\": container with ID starting with d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53 not found: ID does not exist" containerID="d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.456525 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53"} err="failed to get container status \"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53\": rpc error: code = NotFound desc = could not find container \"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53\": container with ID starting with d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53 not found: ID does not exist" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.456609 4837 scope.go:117] "RemoveContainer" containerID="316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.456967 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1"} err="failed to get container status \"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1\": rpc error: code = NotFound desc = could not find container \"316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1\": container with ID starting with 316cc0d22af10b809afead73dabd0952fd26e457fef95e4dc6cdeb4c5d1d45c1 not found: ID does not exist" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.457087 4837 scope.go:117] "RemoveContainer" containerID="d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.457431 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53"} err="failed to get container status \"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53\": rpc error: code = NotFound desc = could not find container \"d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53\": container with ID starting with d9546cb01b7d76c561b1d87b8124f5c57eceed95fe0be59b4a29a9b60f8d6e53 not found: ID does not exist" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.464265 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:35 crc kubenswrapper[4837]: E1001 09:04:35.465305 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.465323 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api" Oct 01 09:04:35 crc kubenswrapper[4837]: E1001 09:04:35.465342 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api-log" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.465348 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api-log" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.465820 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api-log" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.465859 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" containerName="cinder-api" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.468193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.471128 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v2k57" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.471359 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.471480 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.471772 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.472015 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.472245 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.497667 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.553661 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.553805 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.553894 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.553943 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.553979 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.554027 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.554042 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.554081 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.554162 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbrzk\" (UniqueName: \"kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.654861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.654936 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.654975 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655009 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655046 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655066 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655102 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655155 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbrzk\" (UniqueName: \"kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655192 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.655274 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.657773 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.659725 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.660150 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.660813 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.660875 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.664450 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.666716 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.675571 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbrzk\" (UniqueName: \"kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk\") pod \"cinder-api-0\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " pod="openstack/cinder-api-0" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.837514 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e43ee6b-ea30-446f-9c87-8bc8b319423e" path="/var/lib/kubelet/pods/7e43ee6b-ea30-446f-9c87-8bc8b319423e/volumes" Oct 01 09:04:35 crc kubenswrapper[4837]: I1001 09:04:35.839176 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:04:36 crc kubenswrapper[4837]: I1001 09:04:36.131485 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:04:36 crc kubenswrapper[4837]: W1001 09:04:36.135733 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda95b690_f5fd_4c66_8e89_b4d13d5c983b.slice/crio-52a09ec61f6774577b3fe7f1894194c06aa3bde3366781bbccfed8b658819aba WatchSource:0}: Error finding container 52a09ec61f6774577b3fe7f1894194c06aa3bde3366781bbccfed8b658819aba: Status 404 returned error can't find the container with id 52a09ec61f6774577b3fe7f1894194c06aa3bde3366781bbccfed8b658819aba Oct 01 09:04:36 crc kubenswrapper[4837]: I1001 09:04:36.401296 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerStarted","Data":"52a09ec61f6774577b3fe7f1894194c06aa3bde3366781bbccfed8b658819aba"} Oct 01 09:04:37 crc kubenswrapper[4837]: I1001 09:04:37.418709 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerStarted","Data":"06402a941554741fdb4097c872a22bf093bc86405a963169bc733b4f6c1ce051"} Oct 01 09:04:37 crc kubenswrapper[4837]: I1001 09:04:37.419508 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerStarted","Data":"89955ce059a7e135e735784ffc0d60ae4ec4ecca7b792eacd5ea6447fff9fd8d"} Oct 01 09:04:37 crc kubenswrapper[4837]: I1001 09:04:37.419534 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 09:04:37 crc kubenswrapper[4837]: I1001 09:04:37.441585 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.441561504 podStartE2EDuration="2.441561504s" podCreationTimestamp="2025-10-01 09:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:04:37.439287019 +0000 UTC m=+7134.280894514" watchObservedRunningTime="2025-10-01 09:04:37.441561504 +0000 UTC m=+7134.283168959" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.021857 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.090084 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.090474 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="dnsmasq-dns" containerID="cri-o://723bbf703ed1d28b0c5040eb3899764766b5b06150afd86006d625d0e08ccd2c" gracePeriod=10 Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.477620 4837 generic.go:334] "Generic (PLEG): container finished" podID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerID="723bbf703ed1d28b0c5040eb3899764766b5b06150afd86006d625d0e08ccd2c" exitCode=0 Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.477738 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" event={"ID":"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0","Type":"ContainerDied","Data":"723bbf703ed1d28b0c5040eb3899764766b5b06150afd86006d625d0e08ccd2c"} Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.623817 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.679684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb\") pod \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.679767 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvfvd\" (UniqueName: \"kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd\") pod \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.679821 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config\") pod \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.679897 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb\") pod \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.679927 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc\") pod \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\" (UID: \"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0\") " Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.699007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd" (OuterVolumeSpecName: "kube-api-access-nvfvd") pod "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" (UID: "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0"). InnerVolumeSpecName "kube-api-access-nvfvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.732151 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" (UID: "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.733231 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" (UID: "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.735794 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" (UID: "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.747023 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config" (OuterVolumeSpecName: "config") pod "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" (UID: "f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.781917 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.781965 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvfvd\" (UniqueName: \"kubernetes.io/projected/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-kube-api-access-nvfvd\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.781980 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.781993 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.782005 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:04:41 crc kubenswrapper[4837]: I1001 09:04:41.815738 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:04:41 crc kubenswrapper[4837]: E1001 09:04:41.816142 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.496863 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" event={"ID":"f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0","Type":"ContainerDied","Data":"e6db9d86e7fca29c3af63feb82b2dd8d6b36112fbbf002864a8afc84441a2177"} Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.496930 4837 scope.go:117] "RemoveContainer" containerID="723bbf703ed1d28b0c5040eb3899764766b5b06150afd86006d625d0e08ccd2c" Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.496989 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6669cbdcf7-pxd6c" Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.525928 4837 scope.go:117] "RemoveContainer" containerID="8a5b24f86fc0d377779188d082a98168656b387ccd435b2ebc5fa92745516fe1" Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.540010 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:04:42 crc kubenswrapper[4837]: I1001 09:04:42.551434 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6669cbdcf7-pxd6c"] Oct 01 09:04:43 crc kubenswrapper[4837]: I1001 09:04:43.836498 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" path="/var/lib/kubelet/pods/f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0/volumes" Oct 01 09:04:47 crc kubenswrapper[4837]: I1001 09:04:47.747291 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 09:04:56 crc kubenswrapper[4837]: I1001 09:04:56.816402 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:04:56 crc kubenswrapper[4837]: E1001 09:04:56.817353 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:05:06 crc kubenswrapper[4837]: I1001 09:05:06.893883 4837 scope.go:117] "RemoveContainer" containerID="66169f6d97fb753d6c66b2b50344b91e550064a07cd0436defb4dd4fdca8c210" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.651935 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:09 crc kubenswrapper[4837]: E1001 09:05:09.652930 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="dnsmasq-dns" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.652950 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="dnsmasq-dns" Oct 01 09:05:09 crc kubenswrapper[4837]: E1001 09:05:09.652995 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="init" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.653004 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="init" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.653203 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f99ca47f-4a95-4a1f-8342-ff5e2d6af2b0" containerName="dnsmasq-dns" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.654442 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.657818 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.663897 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.682937 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.683021 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.683041 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.683082 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.683131 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slczh\" (UniqueName: \"kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.683205 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.784615 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785111 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785153 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785199 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785244 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slczh\" (UniqueName: \"kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785319 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.785400 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.791213 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.791263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.791350 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.795419 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.806519 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slczh\" (UniqueName: \"kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh\") pod \"cinder-scheduler-0\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:09 crc kubenswrapper[4837]: I1001 09:05:09.977749 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:10 crc kubenswrapper[4837]: I1001 09:05:10.450632 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:10 crc kubenswrapper[4837]: I1001 09:05:10.783478 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerStarted","Data":"10e38b5116f8689958f007e369d5216e5dee15afa9b907e49762930b71dcb163"} Oct 01 09:05:10 crc kubenswrapper[4837]: I1001 09:05:10.818126 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:05:10 crc kubenswrapper[4837]: E1001 09:05:10.818601 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.291673 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.291955 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api-log" containerID="cri-o://89955ce059a7e135e735784ffc0d60ae4ec4ecca7b792eacd5ea6447fff9fd8d" gracePeriod=30 Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.292068 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api" containerID="cri-o://06402a941554741fdb4097c872a22bf093bc86405a963169bc733b4f6c1ce051" gracePeriod=30 Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.797271 4837 generic.go:334] "Generic (PLEG): container finished" podID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerID="89955ce059a7e135e735784ffc0d60ae4ec4ecca7b792eacd5ea6447fff9fd8d" exitCode=143 Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.797492 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerDied","Data":"89955ce059a7e135e735784ffc0d60ae4ec4ecca7b792eacd5ea6447fff9fd8d"} Oct 01 09:05:11 crc kubenswrapper[4837]: I1001 09:05:11.798766 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerStarted","Data":"a167d4e179f13cf7641b455adef11ceafbedcffb22ee26b1d271a980f92832ce"} Oct 01 09:05:12 crc kubenswrapper[4837]: I1001 09:05:12.810342 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerStarted","Data":"e740f074ff84750272ebbf9ec6abeed13e389625d3cc27dc40991c639d9c9412"} Oct 01 09:05:12 crc kubenswrapper[4837]: I1001 09:05:12.834670 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.5138296970000003 podStartE2EDuration="3.834647458s" podCreationTimestamp="2025-10-01 09:05:09 +0000 UTC" firstStartedPulling="2025-10-01 09:05:10.464526326 +0000 UTC m=+7167.306133791" lastFinishedPulling="2025-10-01 09:05:10.785344087 +0000 UTC m=+7167.626951552" observedRunningTime="2025-10-01 09:05:12.831113151 +0000 UTC m=+7169.672720646" watchObservedRunningTime="2025-10-01 09:05:12.834647458 +0000 UTC m=+7169.676254923" Oct 01 09:05:14 crc kubenswrapper[4837]: I1001 09:05:14.837261 4837 generic.go:334] "Generic (PLEG): container finished" podID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerID="06402a941554741fdb4097c872a22bf093bc86405a963169bc733b4f6c1ce051" exitCode=0 Oct 01 09:05:14 crc kubenswrapper[4837]: I1001 09:05:14.837563 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerDied","Data":"06402a941554741fdb4097c872a22bf093bc86405a963169bc733b4f6c1ce051"} Oct 01 09:05:14 crc kubenswrapper[4837]: I1001 09:05:14.925336 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:05:14 crc kubenswrapper[4837]: I1001 09:05:14.978416 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078556 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078624 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078745 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbrzk\" (UniqueName: \"kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078772 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078850 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078873 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078898 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078918 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.078966 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom\") pod \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\" (UID: \"da95b690-f5fd-4c66-8e89-b4d13d5c983b\") " Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.080250 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.081636 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs" (OuterVolumeSpecName: "logs") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.086707 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts" (OuterVolumeSpecName: "scripts") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.086888 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk" (OuterVolumeSpecName: "kube-api-access-dbrzk") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "kube-api-access-dbrzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.102571 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.131827 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.155192 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data" (OuterVolumeSpecName: "config-data") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.157050 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.163478 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "da95b690-f5fd-4c66-8e89-b4d13d5c983b" (UID: "da95b690-f5fd-4c66-8e89-b4d13d5c983b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.180952 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.180984 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da95b690-f5fd-4c66-8e89-b4d13d5c983b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.180993 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181003 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbrzk\" (UniqueName: \"kubernetes.io/projected/da95b690-f5fd-4c66-8e89-b4d13d5c983b-kube-api-access-dbrzk\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181012 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181022 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181030 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da95b690-f5fd-4c66-8e89-b4d13d5c983b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181040 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.181050 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da95b690-f5fd-4c66-8e89-b4d13d5c983b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.850938 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da95b690-f5fd-4c66-8e89-b4d13d5c983b","Type":"ContainerDied","Data":"52a09ec61f6774577b3fe7f1894194c06aa3bde3366781bbccfed8b658819aba"} Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.851023 4837 scope.go:117] "RemoveContainer" containerID="06402a941554741fdb4097c872a22bf093bc86405a963169bc733b4f6c1ce051" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.851196 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.882611 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.902904 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.903143 4837 scope.go:117] "RemoveContainer" containerID="89955ce059a7e135e735784ffc0d60ae4ec4ecca7b792eacd5ea6447fff9fd8d" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.913955 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:15 crc kubenswrapper[4837]: E1001 09:05:15.915095 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api-log" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.915138 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api-log" Oct 01 09:05:15 crc kubenswrapper[4837]: E1001 09:05:15.915193 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.915209 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.915583 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api-log" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.915618 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" containerName="cinder-api" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.919574 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.922770 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.923027 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.923249 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 09:05:15 crc kubenswrapper[4837]: I1001 09:05:15.923854 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000110 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-scripts\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000158 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000180 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000206 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlzn5\" (UniqueName: \"kubernetes.io/projected/c455811f-4366-4ae6-8814-58b17ce42597-kube-api-access-wlzn5\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000546 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c455811f-4366-4ae6-8814-58b17ce42597-logs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000729 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000791 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c455811f-4366-4ae6-8814-58b17ce42597-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000822 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data-custom\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.000851 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102492 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data-custom\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102554 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102622 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-scripts\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102649 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102682 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102761 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlzn5\" (UniqueName: \"kubernetes.io/projected/c455811f-4366-4ae6-8814-58b17ce42597-kube-api-access-wlzn5\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102877 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c455811f-4366-4ae6-8814-58b17ce42597-logs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102946 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.102990 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c455811f-4366-4ae6-8814-58b17ce42597-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.103091 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c455811f-4366-4ae6-8814-58b17ce42597-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.103561 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c455811f-4366-4ae6-8814-58b17ce42597-logs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.109054 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.109084 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-scripts\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.109432 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data-custom\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.110349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-config-data\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.117757 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.121483 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c455811f-4366-4ae6-8814-58b17ce42597-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.131304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlzn5\" (UniqueName: \"kubernetes.io/projected/c455811f-4366-4ae6-8814-58b17ce42597-kube-api-access-wlzn5\") pod \"cinder-api-0\" (UID: \"c455811f-4366-4ae6-8814-58b17ce42597\") " pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.251022 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.548469 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:05:16 crc kubenswrapper[4837]: I1001 09:05:16.868005 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c455811f-4366-4ae6-8814-58b17ce42597","Type":"ContainerStarted","Data":"07ee88040ebb0ec727509020bc5b7dd83639d4a32b1b5dd030a2e08eec32fcff"} Oct 01 09:05:17 crc kubenswrapper[4837]: I1001 09:05:17.838647 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da95b690-f5fd-4c66-8e89-b4d13d5c983b" path="/var/lib/kubelet/pods/da95b690-f5fd-4c66-8e89-b4d13d5c983b/volumes" Oct 01 09:05:17 crc kubenswrapper[4837]: I1001 09:05:17.884286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c455811f-4366-4ae6-8814-58b17ce42597","Type":"ContainerStarted","Data":"f4d2c3964272e0ba1b13b1f776772593be3ec014c1fb17d93123efccca850205"} Oct 01 09:05:18 crc kubenswrapper[4837]: I1001 09:05:18.896815 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c455811f-4366-4ae6-8814-58b17ce42597","Type":"ContainerStarted","Data":"9960c40aa66f3006dc4939a35b3453c0a78d59e8bd587ad12eff15adec5f2f62"} Oct 01 09:05:18 crc kubenswrapper[4837]: I1001 09:05:18.897724 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 09:05:18 crc kubenswrapper[4837]: I1001 09:05:18.942309 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.942289021 podStartE2EDuration="3.942289021s" podCreationTimestamp="2025-10-01 09:05:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:05:18.929071686 +0000 UTC m=+7175.770679161" watchObservedRunningTime="2025-10-01 09:05:18.942289021 +0000 UTC m=+7175.783896486" Oct 01 09:05:20 crc kubenswrapper[4837]: I1001 09:05:20.225243 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 09:05:20 crc kubenswrapper[4837]: I1001 09:05:20.265938 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:20 crc kubenswrapper[4837]: I1001 09:05:20.920385 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="cinder-scheduler" containerID="cri-o://a167d4e179f13cf7641b455adef11ceafbedcffb22ee26b1d271a980f92832ce" gracePeriod=30 Oct 01 09:05:20 crc kubenswrapper[4837]: I1001 09:05:20.920464 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="probe" containerID="cri-o://e740f074ff84750272ebbf9ec6abeed13e389625d3cc27dc40991c639d9c9412" gracePeriod=30 Oct 01 09:05:21 crc kubenswrapper[4837]: I1001 09:05:21.932667 4837 generic.go:334] "Generic (PLEG): container finished" podID="152aa985-da78-408e-86fa-08f21d59112f" containerID="e740f074ff84750272ebbf9ec6abeed13e389625d3cc27dc40991c639d9c9412" exitCode=0 Oct 01 09:05:21 crc kubenswrapper[4837]: I1001 09:05:21.932742 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerDied","Data":"e740f074ff84750272ebbf9ec6abeed13e389625d3cc27dc40991c639d9c9412"} Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.816822 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:05:22 crc kubenswrapper[4837]: E1001 09:05:22.817303 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.948068 4837 generic.go:334] "Generic (PLEG): container finished" podID="152aa985-da78-408e-86fa-08f21d59112f" containerID="a167d4e179f13cf7641b455adef11ceafbedcffb22ee26b1d271a980f92832ce" exitCode=0 Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.948143 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerDied","Data":"a167d4e179f13cf7641b455adef11ceafbedcffb22ee26b1d271a980f92832ce"} Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.948192 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"152aa985-da78-408e-86fa-08f21d59112f","Type":"ContainerDied","Data":"10e38b5116f8689958f007e369d5216e5dee15afa9b907e49762930b71dcb163"} Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.948215 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10e38b5116f8689958f007e369d5216e5dee15afa9b907e49762930b71dcb163" Oct 01 09:05:22 crc kubenswrapper[4837]: I1001 09:05:22.989497 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.158310 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.158951 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.158998 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.159052 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.159082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.159130 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.159164 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slczh\" (UniqueName: \"kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh\") pod \"152aa985-da78-408e-86fa-08f21d59112f\" (UID: \"152aa985-da78-408e-86fa-08f21d59112f\") " Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.159626 4837 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/152aa985-da78-408e-86fa-08f21d59112f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.165107 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.165271 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts" (OuterVolumeSpecName: "scripts") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.184191 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh" (OuterVolumeSpecName: "kube-api-access-slczh") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "kube-api-access-slczh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.214835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.255426 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data" (OuterVolumeSpecName: "config-data") pod "152aa985-da78-408e-86fa-08f21d59112f" (UID: "152aa985-da78-408e-86fa-08f21d59112f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.261216 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.261249 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.261261 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slczh\" (UniqueName: \"kubernetes.io/projected/152aa985-da78-408e-86fa-08f21d59112f-kube-api-access-slczh\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.261270 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.261278 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/152aa985-da78-408e-86fa-08f21d59112f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.959229 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:23 crc kubenswrapper[4837]: I1001 09:05:23.998472 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.007986 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.017398 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:24 crc kubenswrapper[4837]: E1001 09:05:24.017819 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="probe" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.017840 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="probe" Oct 01 09:05:24 crc kubenswrapper[4837]: E1001 09:05:24.017906 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="cinder-scheduler" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.017916 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="cinder-scheduler" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.018158 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="cinder-scheduler" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.018192 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="152aa985-da78-408e-86fa-08f21d59112f" containerName="probe" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.019324 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.021661 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.039177 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190054 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-scripts\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190101 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190213 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl4pk\" (UniqueName: \"kubernetes.io/projected/3619414f-7707-4a54-a834-bf1a1f346a14-kube-api-access-bl4pk\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3619414f-7707-4a54-a834-bf1a1f346a14-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.190579 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292187 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292263 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-scripts\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292307 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292391 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl4pk\" (UniqueName: \"kubernetes.io/projected/3619414f-7707-4a54-a834-bf1a1f346a14-kube-api-access-bl4pk\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3619414f-7707-4a54-a834-bf1a1f346a14-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.292676 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3619414f-7707-4a54-a834-bf1a1f346a14-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.299631 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.305369 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.305542 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.305742 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3619414f-7707-4a54-a834-bf1a1f346a14-scripts\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.317874 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl4pk\" (UniqueName: \"kubernetes.io/projected/3619414f-7707-4a54-a834-bf1a1f346a14-kube-api-access-bl4pk\") pod \"cinder-scheduler-0\" (UID: \"3619414f-7707-4a54-a834-bf1a1f346a14\") " pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.347832 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.867658 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:05:24 crc kubenswrapper[4837]: I1001 09:05:24.968823 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3619414f-7707-4a54-a834-bf1a1f346a14","Type":"ContainerStarted","Data":"f0893a575a4194e3a49a9b6f50b2442d1d1d8d5144a7339eb30c80ed1f9e1425"} Oct 01 09:05:25 crc kubenswrapper[4837]: I1001 09:05:25.831871 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="152aa985-da78-408e-86fa-08f21d59112f" path="/var/lib/kubelet/pods/152aa985-da78-408e-86fa-08f21d59112f/volumes" Oct 01 09:05:25 crc kubenswrapper[4837]: I1001 09:05:25.981637 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3619414f-7707-4a54-a834-bf1a1f346a14","Type":"ContainerStarted","Data":"66d68ff2ae2c1fa811ee28f1ebc1be6f8cdcd425ff3ffc158ab924ba10c8b7e3"} Oct 01 09:05:26 crc kubenswrapper[4837]: I1001 09:05:26.993834 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3619414f-7707-4a54-a834-bf1a1f346a14","Type":"ContainerStarted","Data":"141bd23adf672fd3b88dad3b872f2da8d6a3238e2dc2276ed5b6cc0077980551"} Oct 01 09:05:27 crc kubenswrapper[4837]: I1001 09:05:27.025976 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.025952864 podStartE2EDuration="4.025952864s" podCreationTimestamp="2025-10-01 09:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:05:27.019223178 +0000 UTC m=+7183.860830633" watchObservedRunningTime="2025-10-01 09:05:27.025952864 +0000 UTC m=+7183.867560319" Oct 01 09:05:28 crc kubenswrapper[4837]: I1001 09:05:28.023981 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 09:05:29 crc kubenswrapper[4837]: I1001 09:05:29.348849 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 09:05:34 crc kubenswrapper[4837]: I1001 09:05:34.602296 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 09:05:35 crc kubenswrapper[4837]: I1001 09:05:35.984961 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6vtpd"] Oct 01 09:05:35 crc kubenswrapper[4837]: I1001 09:05:35.986513 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:35 crc kubenswrapper[4837]: I1001 09:05:35.999322 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6vtpd"] Oct 01 09:05:36 crc kubenswrapper[4837]: I1001 09:05:36.140923 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn64v\" (UniqueName: \"kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v\") pod \"glance-db-create-6vtpd\" (UID: \"9823550a-12f9-4827-a8cc-8edb88305a2e\") " pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:36 crc kubenswrapper[4837]: I1001 09:05:36.242752 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn64v\" (UniqueName: \"kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v\") pod \"glance-db-create-6vtpd\" (UID: \"9823550a-12f9-4827-a8cc-8edb88305a2e\") " pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:36 crc kubenswrapper[4837]: I1001 09:05:36.273601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn64v\" (UniqueName: \"kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v\") pod \"glance-db-create-6vtpd\" (UID: \"9823550a-12f9-4827-a8cc-8edb88305a2e\") " pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:36 crc kubenswrapper[4837]: I1001 09:05:36.309627 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:36 crc kubenswrapper[4837]: I1001 09:05:36.855761 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6vtpd"] Oct 01 09:05:37 crc kubenswrapper[4837]: I1001 09:05:37.117337 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6vtpd" event={"ID":"9823550a-12f9-4827-a8cc-8edb88305a2e","Type":"ContainerStarted","Data":"1b4c3eb4aaa4f6a612aecf21a42ce7ae1a340a4c6e4037e04d5a5589c3c2f3ec"} Oct 01 09:05:37 crc kubenswrapper[4837]: I1001 09:05:37.142870 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-6vtpd" podStartSLOduration=2.142846584 podStartE2EDuration="2.142846584s" podCreationTimestamp="2025-10-01 09:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:05:37.135524384 +0000 UTC m=+7193.977131839" watchObservedRunningTime="2025-10-01 09:05:37.142846584 +0000 UTC m=+7193.984454069" Oct 01 09:05:37 crc kubenswrapper[4837]: I1001 09:05:37.816141 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:05:37 crc kubenswrapper[4837]: E1001 09:05:37.816423 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:05:38 crc kubenswrapper[4837]: I1001 09:05:38.132553 4837 generic.go:334] "Generic (PLEG): container finished" podID="9823550a-12f9-4827-a8cc-8edb88305a2e" containerID="be001c6b3805e369ac05561f5c7aca7952a4dd023f0db6bcc76af9aa898e549f" exitCode=0 Oct 01 09:05:38 crc kubenswrapper[4837]: I1001 09:05:38.132726 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6vtpd" event={"ID":"9823550a-12f9-4827-a8cc-8edb88305a2e","Type":"ContainerDied","Data":"be001c6b3805e369ac05561f5c7aca7952a4dd023f0db6bcc76af9aa898e549f"} Oct 01 09:05:39 crc kubenswrapper[4837]: I1001 09:05:39.562427 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:39 crc kubenswrapper[4837]: I1001 09:05:39.715346 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn64v\" (UniqueName: \"kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v\") pod \"9823550a-12f9-4827-a8cc-8edb88305a2e\" (UID: \"9823550a-12f9-4827-a8cc-8edb88305a2e\") " Oct 01 09:05:39 crc kubenswrapper[4837]: I1001 09:05:39.724116 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v" (OuterVolumeSpecName: "kube-api-access-mn64v") pod "9823550a-12f9-4827-a8cc-8edb88305a2e" (UID: "9823550a-12f9-4827-a8cc-8edb88305a2e"). InnerVolumeSpecName "kube-api-access-mn64v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:05:39 crc kubenswrapper[4837]: I1001 09:05:39.819173 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn64v\" (UniqueName: \"kubernetes.io/projected/9823550a-12f9-4827-a8cc-8edb88305a2e-kube-api-access-mn64v\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:40 crc kubenswrapper[4837]: I1001 09:05:40.160077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6vtpd" event={"ID":"9823550a-12f9-4827-a8cc-8edb88305a2e","Type":"ContainerDied","Data":"1b4c3eb4aaa4f6a612aecf21a42ce7ae1a340a4c6e4037e04d5a5589c3c2f3ec"} Oct 01 09:05:40 crc kubenswrapper[4837]: I1001 09:05:40.160185 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b4c3eb4aaa4f6a612aecf21a42ce7ae1a340a4c6e4037e04d5a5589c3c2f3ec" Oct 01 09:05:40 crc kubenswrapper[4837]: I1001 09:05:40.160301 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6vtpd" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.127159 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-7f7c-account-create-w8s49"] Oct 01 09:05:46 crc kubenswrapper[4837]: E1001 09:05:46.128934 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9823550a-12f9-4827-a8cc-8edb88305a2e" containerName="mariadb-database-create" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.128972 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9823550a-12f9-4827-a8cc-8edb88305a2e" containerName="mariadb-database-create" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.129326 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9823550a-12f9-4827-a8cc-8edb88305a2e" containerName="mariadb-database-create" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.130548 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.134792 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.139293 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7f7c-account-create-w8s49"] Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.261126 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbb9z\" (UniqueName: \"kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z\") pod \"glance-7f7c-account-create-w8s49\" (UID: \"9ca515d7-b285-42a1-9759-6e68115bffbf\") " pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.363463 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbb9z\" (UniqueName: \"kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z\") pod \"glance-7f7c-account-create-w8s49\" (UID: \"9ca515d7-b285-42a1-9759-6e68115bffbf\") " pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.391314 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbb9z\" (UniqueName: \"kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z\") pod \"glance-7f7c-account-create-w8s49\" (UID: \"9ca515d7-b285-42a1-9759-6e68115bffbf\") " pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.464553 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:46 crc kubenswrapper[4837]: I1001 09:05:46.788008 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7f7c-account-create-w8s49"] Oct 01 09:05:46 crc kubenswrapper[4837]: W1001 09:05:46.801523 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ca515d7_b285_42a1_9759_6e68115bffbf.slice/crio-016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1 WatchSource:0}: Error finding container 016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1: Status 404 returned error can't find the container with id 016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1 Oct 01 09:05:47 crc kubenswrapper[4837]: I1001 09:05:47.248630 4837 generic.go:334] "Generic (PLEG): container finished" podID="9ca515d7-b285-42a1-9759-6e68115bffbf" containerID="beaf1bb636682fcc43dfa32c437449a18f700ec677d5e2f1163a677fe7031673" exitCode=0 Oct 01 09:05:47 crc kubenswrapper[4837]: I1001 09:05:47.248772 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7f7c-account-create-w8s49" event={"ID":"9ca515d7-b285-42a1-9759-6e68115bffbf","Type":"ContainerDied","Data":"beaf1bb636682fcc43dfa32c437449a18f700ec677d5e2f1163a677fe7031673"} Oct 01 09:05:47 crc kubenswrapper[4837]: I1001 09:05:47.248900 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7f7c-account-create-w8s49" event={"ID":"9ca515d7-b285-42a1-9759-6e68115bffbf","Type":"ContainerStarted","Data":"016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1"} Oct 01 09:05:48 crc kubenswrapper[4837]: I1001 09:05:48.671902 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:48 crc kubenswrapper[4837]: I1001 09:05:48.711526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbb9z\" (UniqueName: \"kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z\") pod \"9ca515d7-b285-42a1-9759-6e68115bffbf\" (UID: \"9ca515d7-b285-42a1-9759-6e68115bffbf\") " Oct 01 09:05:48 crc kubenswrapper[4837]: I1001 09:05:48.719049 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z" (OuterVolumeSpecName: "kube-api-access-fbb9z") pod "9ca515d7-b285-42a1-9759-6e68115bffbf" (UID: "9ca515d7-b285-42a1-9759-6e68115bffbf"). InnerVolumeSpecName "kube-api-access-fbb9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:05:48 crc kubenswrapper[4837]: I1001 09:05:48.813803 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbb9z\" (UniqueName: \"kubernetes.io/projected/9ca515d7-b285-42a1-9759-6e68115bffbf-kube-api-access-fbb9z\") on node \"crc\" DevicePath \"\"" Oct 01 09:05:49 crc kubenswrapper[4837]: I1001 09:05:49.274296 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7f7c-account-create-w8s49" event={"ID":"9ca515d7-b285-42a1-9759-6e68115bffbf","Type":"ContainerDied","Data":"016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1"} Oct 01 09:05:49 crc kubenswrapper[4837]: I1001 09:05:49.274368 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="016fb8a5e89514321c3866988f27501cf31370bc7ff9ee661083df3f8fd5cfa1" Oct 01 09:05:49 crc kubenswrapper[4837]: I1001 09:05:49.274481 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7f7c-account-create-w8s49" Oct 01 09:05:49 crc kubenswrapper[4837]: I1001 09:05:49.816826 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:05:49 crc kubenswrapper[4837]: E1001 09:05:49.817412 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.192587 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-f2wtt"] Oct 01 09:05:51 crc kubenswrapper[4837]: E1001 09:05:51.193581 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca515d7-b285-42a1-9759-6e68115bffbf" containerName="mariadb-account-create" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.193603 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca515d7-b285-42a1-9759-6e68115bffbf" containerName="mariadb-account-create" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.193952 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ca515d7-b285-42a1-9759-6e68115bffbf" containerName="mariadb-account-create" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.195665 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.199190 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.199312 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-swd2f" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.208145 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-f2wtt"] Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.365919 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.366019 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.366100 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksxvv\" (UniqueName: \"kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.366325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.467878 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.467992 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.468008 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.468041 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksxvv\" (UniqueName: \"kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.476587 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.477102 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.481187 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.491418 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksxvv\" (UniqueName: \"kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv\") pod \"glance-db-sync-f2wtt\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:51 crc kubenswrapper[4837]: I1001 09:05:51.523176 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-f2wtt" Oct 01 09:05:52 crc kubenswrapper[4837]: I1001 09:05:52.087990 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-f2wtt"] Oct 01 09:05:52 crc kubenswrapper[4837]: I1001 09:05:52.309195 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-f2wtt" event={"ID":"0bf0317f-dd4d-40e8-aa61-5e22923f51f0","Type":"ContainerStarted","Data":"4da2d68f9ab5471834a4eb867ad8f01ba2dbc60c21e7e15cbf13d9d02c496765"} Oct 01 09:06:00 crc kubenswrapper[4837]: I1001 09:06:00.816607 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:06:00 crc kubenswrapper[4837]: E1001 09:06:00.817526 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.626766 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.629809 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.633687 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.764901 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.765471 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.765757 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f58l\" (UniqueName: \"kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.874311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f58l\" (UniqueName: \"kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.874457 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.875138 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.875229 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.875555 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.899849 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f58l\" (UniqueName: \"kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l\") pod \"community-operators-2d7sd\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:06 crc kubenswrapper[4837]: I1001 09:06:06.986666 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.075144 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.488268 4837 generic.go:334] "Generic (PLEG): container finished" podID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerID="6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce" exitCode=0 Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.488406 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerDied","Data":"6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce"} Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.488620 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerStarted","Data":"701c37c59ad64b33c471758de80213b593d97fe1912ab50d8aaec6ee7d3511bd"} Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.490570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-f2wtt" event={"ID":"0bf0317f-dd4d-40e8-aa61-5e22923f51f0","Type":"ContainerStarted","Data":"722355f307a5bd3e6b2812b6a2eb07fa4571e3767b5a480aec16291c47a68e43"} Oct 01 09:06:10 crc kubenswrapper[4837]: I1001 09:06:10.528380 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-f2wtt" podStartSLOduration=1.949001097 podStartE2EDuration="19.528348171s" podCreationTimestamp="2025-10-01 09:05:51 +0000 UTC" firstStartedPulling="2025-10-01 09:05:52.095760608 +0000 UTC m=+7208.937368063" lastFinishedPulling="2025-10-01 09:06:09.675107642 +0000 UTC m=+7226.516715137" observedRunningTime="2025-10-01 09:06:10.524041325 +0000 UTC m=+7227.365648780" watchObservedRunningTime="2025-10-01 09:06:10.528348171 +0000 UTC m=+7227.369955636" Oct 01 09:06:12 crc kubenswrapper[4837]: I1001 09:06:12.529860 4837 generic.go:334] "Generic (PLEG): container finished" podID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerID="e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a" exitCode=0 Oct 01 09:06:12 crc kubenswrapper[4837]: I1001 09:06:12.529918 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerDied","Data":"e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a"} Oct 01 09:06:13 crc kubenswrapper[4837]: I1001 09:06:13.540249 4837 generic.go:334] "Generic (PLEG): container finished" podID="0bf0317f-dd4d-40e8-aa61-5e22923f51f0" containerID="722355f307a5bd3e6b2812b6a2eb07fa4571e3767b5a480aec16291c47a68e43" exitCode=0 Oct 01 09:06:13 crc kubenswrapper[4837]: I1001 09:06:13.540373 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-f2wtt" event={"ID":"0bf0317f-dd4d-40e8-aa61-5e22923f51f0","Type":"ContainerDied","Data":"722355f307a5bd3e6b2812b6a2eb07fa4571e3767b5a480aec16291c47a68e43"} Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.552208 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerStarted","Data":"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f"} Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.577679 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2d7sd" podStartSLOduration=5.693721338 podStartE2EDuration="8.577654691s" podCreationTimestamp="2025-10-01 09:06:06 +0000 UTC" firstStartedPulling="2025-10-01 09:06:10.489778434 +0000 UTC m=+7227.331385909" lastFinishedPulling="2025-10-01 09:06:13.373711797 +0000 UTC m=+7230.215319262" observedRunningTime="2025-10-01 09:06:14.572431124 +0000 UTC m=+7231.414038589" watchObservedRunningTime="2025-10-01 09:06:14.577654691 +0000 UTC m=+7231.419262146" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.707013 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.709418 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.719748 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.743548 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.743615 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8lvs\" (UniqueName: \"kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.743998 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.816487 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:06:14 crc kubenswrapper[4837]: E1001 09:06:14.816750 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.846209 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8lvs\" (UniqueName: \"kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.846322 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.846396 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.846792 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.847983 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.866963 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8lvs\" (UniqueName: \"kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs\") pod \"redhat-operators-q2t4n\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.916239 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.935463 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.941932 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.947279 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.947322 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjfpp\" (UniqueName: \"kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.947367 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:14 crc kubenswrapper[4837]: I1001 09:06:14.987747 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-f2wtt" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.041711 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.048555 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle\") pod \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.048677 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data\") pod \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.048751 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksxvv\" (UniqueName: \"kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv\") pod \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.048860 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data\") pod \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\" (UID: \"0bf0317f-dd4d-40e8-aa61-5e22923f51f0\") " Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.049134 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.050113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.050220 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.050555 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.050292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjfpp\" (UniqueName: \"kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.058110 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0bf0317f-dd4d-40e8-aa61-5e22923f51f0" (UID: "0bf0317f-dd4d-40e8-aa61-5e22923f51f0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.067276 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjfpp\" (UniqueName: \"kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp\") pod \"certified-operators-znjtf\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.072001 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv" (OuterVolumeSpecName: "kube-api-access-ksxvv") pod "0bf0317f-dd4d-40e8-aa61-5e22923f51f0" (UID: "0bf0317f-dd4d-40e8-aa61-5e22923f51f0"). InnerVolumeSpecName "kube-api-access-ksxvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.090874 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf0317f-dd4d-40e8-aa61-5e22923f51f0" (UID: "0bf0317f-dd4d-40e8-aa61-5e22923f51f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.105476 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data" (OuterVolumeSpecName: "config-data") pod "0bf0317f-dd4d-40e8-aa61-5e22923f51f0" (UID: "0bf0317f-dd4d-40e8-aa61-5e22923f51f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.152342 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.152376 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.152388 4837 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.152396 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksxvv\" (UniqueName: \"kubernetes.io/projected/0bf0317f-dd4d-40e8-aa61-5e22923f51f0-kube-api-access-ksxvv\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.298621 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.557159 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.591081 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-f2wtt" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.591749 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-f2wtt" event={"ID":"0bf0317f-dd4d-40e8-aa61-5e22923f51f0","Type":"ContainerDied","Data":"4da2d68f9ab5471834a4eb867ad8f01ba2dbc60c21e7e15cbf13d9d02c496765"} Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.591781 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4da2d68f9ab5471834a4eb867ad8f01ba2dbc60c21e7e15cbf13d9d02c496765" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.627251 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.886705 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:15 crc kubenswrapper[4837]: E1001 09:06:15.887296 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf0317f-dd4d-40e8-aa61-5e22923f51f0" containerName="glance-db-sync" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.887312 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf0317f-dd4d-40e8-aa61-5e22923f51f0" containerName="glance-db-sync" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.887463 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf0317f-dd4d-40e8-aa61-5e22923f51f0" containerName="glance-db-sync" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.888405 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.894895 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.895199 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-swd2f" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.895312 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.905501 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.973718 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.973995 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.974159 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.974283 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.974419 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cthc9\" (UniqueName: \"kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:15 crc kubenswrapper[4837]: I1001 09:06:15.974536 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.005004 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.007422 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.022108 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.075958 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076017 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076074 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076099 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076117 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076148 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076174 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cthc9\" (UniqueName: \"kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076194 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076223 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076239 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.076253 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpj2t\" (UniqueName: \"kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.081217 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.081304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.091558 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.092553 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.093318 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.093712 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.104140 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.104477 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.128181 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.129465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cthc9\" (UniqueName: \"kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9\") pod \"glance-default-external-api-0\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.177607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.177680 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181752 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181797 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181854 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181898 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181921 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpj2t\" (UniqueName: \"kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181945 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.181999 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.182063 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.182111 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l25j\" (UniqueName: \"kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.183173 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.183922 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.184603 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.185514 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.226403 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpj2t\" (UniqueName: \"kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t\") pod \"dnsmasq-dns-7c9cf6db59-ts944\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283060 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283110 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283146 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283214 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l25j\" (UniqueName: \"kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283254 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283272 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.283813 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.284031 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.287186 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.291314 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.291768 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.301431 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l25j\" (UniqueName: \"kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j\") pod \"glance-default-internal-api-0\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.326597 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.355089 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.566141 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.645143 4837 generic.go:334] "Generic (PLEG): container finished" podID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerID="6cc321bee8f13b9f7a06548504ff31a32125034c0e7f22010712b6592bb31b9a" exitCode=0 Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.645243 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerDied","Data":"6cc321bee8f13b9f7a06548504ff31a32125034c0e7f22010712b6592bb31b9a"} Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.645271 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerStarted","Data":"cef436345f85ac1e8a8d7c31e4d7feffcc98dc22a14b8a354b2b4b0a595b46bf"} Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.694421 4837 generic.go:334] "Generic (PLEG): container finished" podID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerID="9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d" exitCode=0 Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.694472 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerDied","Data":"9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d"} Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.694501 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerStarted","Data":"46283682e5b0571bd5aeb0ebe8087bdf508a457a7767b2237b11fa00b4b8ffe2"} Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.987158 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:16 crc kubenswrapper[4837]: I1001 09:06:16.987811 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.032677 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.083417 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.087407 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:17 crc kubenswrapper[4837]: W1001 09:06:17.125933 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0913ad8_e8a5_4fd0_94c1_36a6f0b2190b.slice/crio-d0e7754238fc27e029c3977aa70dfe65a538511fd5691cdaa817a039fb403edc WatchSource:0}: Error finding container d0e7754238fc27e029c3977aa70dfe65a538511fd5691cdaa817a039fb403edc: Status 404 returned error can't find the container with id d0e7754238fc27e029c3977aa70dfe65a538511fd5691cdaa817a039fb403edc Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.322641 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:17 crc kubenswrapper[4837]: W1001 09:06:17.325155 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6ce8c6_8685_4205_b2f4_36195acfadaa.slice/crio-1104f6938f83c4ed83f5797cdf74301c2cc0d500392a1ca82572c5761579d860 WatchSource:0}: Error finding container 1104f6938f83c4ed83f5797cdf74301c2cc0d500392a1ca82572c5761579d860: Status 404 returned error can't find the container with id 1104f6938f83c4ed83f5797cdf74301c2cc0d500392a1ca82572c5761579d860 Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.726308 4837 generic.go:334] "Generic (PLEG): container finished" podID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerID="8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e" exitCode=0 Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.726373 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" event={"ID":"829d0b84-e85a-4eb8-a3e5-8bb3502e194f","Type":"ContainerDied","Data":"8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e"} Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.726402 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" event={"ID":"829d0b84-e85a-4eb8-a3e5-8bb3502e194f","Type":"ContainerStarted","Data":"f933b7c398ed4941be951e491759bca3e983c22a4ba4abfb040dd396988a396e"} Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.738249 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerStarted","Data":"d0e7754238fc27e029c3977aa70dfe65a538511fd5691cdaa817a039fb403edc"} Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.741256 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerStarted","Data":"1104f6938f83c4ed83f5797cdf74301c2cc0d500392a1ca82572c5761579d860"} Oct 01 09:06:17 crc kubenswrapper[4837]: I1001 09:06:17.770748 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.422814 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.752329 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerStarted","Data":"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.754659 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" event={"ID":"829d0b84-e85a-4eb8-a3e5-8bb3502e194f","Type":"ContainerStarted","Data":"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.755641 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.758319 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerStarted","Data":"7c682076542a3610bbe65bd23558f10104bc0ccb17908bc4de955087dac16e7b"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.758374 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerStarted","Data":"4c6ee7b9ec34581e58388a51448a539c6885341912eebcad0a4e9636b9d11b61"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.761303 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerStarted","Data":"d312d698953799dcc3220c2a81316e2aa47bf7f308c8939d8bd7ea375ba8909a"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.763417 4837 generic.go:334] "Generic (PLEG): container finished" podID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerID="63b0393d058faa43534c9838df4bc5172697b51db818348385271c07178ae2fd" exitCode=0 Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.763462 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerDied","Data":"63b0393d058faa43534c9838df4bc5172697b51db818348385271c07178ae2fd"} Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.766185 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:06:18 crc kubenswrapper[4837]: I1001 09:06:18.804209 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" podStartSLOduration=3.804190666 podStartE2EDuration="3.804190666s" podCreationTimestamp="2025-10-01 09:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:18.801245793 +0000 UTC m=+7235.642853248" watchObservedRunningTime="2025-10-01 09:06:18.804190666 +0000 UTC m=+7235.645798121" Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.773360 4837 generic.go:334] "Generic (PLEG): container finished" podID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerID="e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db" exitCode=0 Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.773436 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerDied","Data":"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db"} Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.780824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerStarted","Data":"9135880c406079885648b91b41246182264936a00253f59d7504f05ece6dfaa3"} Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.780967 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-log" containerID="cri-o://d312d698953799dcc3220c2a81316e2aa47bf7f308c8939d8bd7ea375ba8909a" gracePeriod=30 Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.781053 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-log" containerID="cri-o://4c6ee7b9ec34581e58388a51448a539c6885341912eebcad0a4e9636b9d11b61" gracePeriod=30 Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.781099 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-httpd" containerID="cri-o://9135880c406079885648b91b41246182264936a00253f59d7504f05ece6dfaa3" gracePeriod=30 Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.781306 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-httpd" containerID="cri-o://7c682076542a3610bbe65bd23558f10104bc0ccb17908bc4de955087dac16e7b" gracePeriod=30 Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.824291 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.824270214 podStartE2EDuration="4.824270214s" podCreationTimestamp="2025-10-01 09:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:19.821211798 +0000 UTC m=+7236.662819273" watchObservedRunningTime="2025-10-01 09:06:19.824270214 +0000 UTC m=+7236.665877679" Oct 01 09:06:19 crc kubenswrapper[4837]: I1001 09:06:19.841985 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.8419552980000002 podStartE2EDuration="3.841955298s" podCreationTimestamp="2025-10-01 09:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:19.84160692 +0000 UTC m=+7236.683214375" watchObservedRunningTime="2025-10-01 09:06:19.841955298 +0000 UTC m=+7236.683562763" Oct 01 09:06:20 crc kubenswrapper[4837]: I1001 09:06:20.796826 4837 generic.go:334] "Generic (PLEG): container finished" podID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerID="4c6ee7b9ec34581e58388a51448a539c6885341912eebcad0a4e9636b9d11b61" exitCode=143 Oct 01 09:06:20 crc kubenswrapper[4837]: I1001 09:06:20.796930 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerDied","Data":"4c6ee7b9ec34581e58388a51448a539c6885341912eebcad0a4e9636b9d11b61"} Oct 01 09:06:20 crc kubenswrapper[4837]: I1001 09:06:20.798538 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerID="d312d698953799dcc3220c2a81316e2aa47bf7f308c8939d8bd7ea375ba8909a" exitCode=143 Oct 01 09:06:20 crc kubenswrapper[4837]: I1001 09:06:20.798680 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerDied","Data":"d312d698953799dcc3220c2a81316e2aa47bf7f308c8939d8bd7ea375ba8909a"} Oct 01 09:06:21 crc kubenswrapper[4837]: I1001 09:06:21.816640 4837 generic.go:334] "Generic (PLEG): container finished" podID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerID="7c682076542a3610bbe65bd23558f10104bc0ccb17908bc4de955087dac16e7b" exitCode=0 Oct 01 09:06:21 crc kubenswrapper[4837]: I1001 09:06:21.823647 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerID="9135880c406079885648b91b41246182264936a00253f59d7504f05ece6dfaa3" exitCode=143 Oct 01 09:06:21 crc kubenswrapper[4837]: I1001 09:06:21.840750 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerDied","Data":"7c682076542a3610bbe65bd23558f10104bc0ccb17908bc4de955087dac16e7b"} Oct 01 09:06:21 crc kubenswrapper[4837]: I1001 09:06:21.841043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerDied","Data":"9135880c406079885648b91b41246182264936a00253f59d7504f05ece6dfaa3"} Oct 01 09:06:21 crc kubenswrapper[4837]: I1001 09:06:21.841195 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerStarted","Data":"28e75f209fdcf6915703e965b75f6118853439623447694211142bbf4a202d0a"} Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.210811 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.311738 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.311812 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.311855 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.311912 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l25j\" (UniqueName: \"kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.311954 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.312229 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data\") pod \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\" (UID: \"7b6ce8c6-8685-4205-b2f4-36195acfadaa\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.313140 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.313316 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs" (OuterVolumeSpecName: "logs") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.318836 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts" (OuterVolumeSpecName: "scripts") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.320543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j" (OuterVolumeSpecName: "kube-api-access-9l25j") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "kube-api-access-9l25j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.346112 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.387609 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data" (OuterVolumeSpecName: "config-data") pod "7b6ce8c6-8685-4205-b2f4-36195acfadaa" (UID: "7b6ce8c6-8685-4205-b2f4-36195acfadaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414000 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414048 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414060 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414077 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b6ce8c6-8685-4205-b2f4-36195acfadaa-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414090 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l25j\" (UniqueName: \"kubernetes.io/projected/7b6ce8c6-8685-4205-b2f4-36195acfadaa-kube-api-access-9l25j\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.414102 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6ce8c6-8685-4205-b2f4-36195acfadaa-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.843342 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7b6ce8c6-8685-4205-b2f4-36195acfadaa","Type":"ContainerDied","Data":"1104f6938f83c4ed83f5797cdf74301c2cc0d500392a1ca82572c5761579d860"} Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.843712 4837 scope.go:117] "RemoveContainer" containerID="9135880c406079885648b91b41246182264936a00253f59d7504f05ece6dfaa3" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.843372 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.872874 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-znjtf" podStartSLOduration=4.58821751 podStartE2EDuration="8.872670627s" podCreationTimestamp="2025-10-01 09:06:14 +0000 UTC" firstStartedPulling="2025-10-01 09:06:16.662996968 +0000 UTC m=+7233.504604423" lastFinishedPulling="2025-10-01 09:06:20.947450055 +0000 UTC m=+7237.789057540" observedRunningTime="2025-10-01 09:06:22.867231153 +0000 UTC m=+7239.708838628" watchObservedRunningTime="2025-10-01 09:06:22.872670627 +0000 UTC m=+7239.714278082" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.903157 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.915450 4837 scope.go:117] "RemoveContainer" containerID="d312d698953799dcc3220c2a81316e2aa47bf7f308c8939d8bd7ea375ba8909a" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924333 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cthc9\" (UniqueName: \"kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924385 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924432 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924532 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.924566 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle\") pod \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\" (UID: \"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b\") " Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.931305 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs" (OuterVolumeSpecName: "logs") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.933467 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.942780 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts" (OuterVolumeSpecName: "scripts") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.951623 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.965052 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9" (OuterVolumeSpecName: "kube-api-access-cthc9") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "kube-api-access-cthc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.981997 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:22 crc kubenswrapper[4837]: I1001 09:06:22.992662 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.011462 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: E1001 09:06:23.011850 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.011865 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: E1001 09:06:23.011880 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.011885 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: E1001 09:06:23.011910 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.011916 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: E1001 09:06:23.011933 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.011939 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.012108 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.012121 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-httpd" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.012130 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.012139 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" containerName="glance-log" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.013193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.014828 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.017846 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034359 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034424 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034442 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034495 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034515 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034562 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pln\" (UniqueName: \"kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034578 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034849 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data" (OuterVolumeSpecName: "config-data") pod "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" (UID: "d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.034930 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.035013 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.035073 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.035129 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.035228 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cthc9\" (UniqueName: \"kubernetes.io/projected/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-kube-api-access-cthc9\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.041283 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136670 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pln\" (UniqueName: \"kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136729 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136775 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136815 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136835 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136885 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136908 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.136962 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.138178 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.138612 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.142896 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.143982 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.145537 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.156034 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pln\" (UniqueName: \"kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.158605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.335280 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.831773 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6ce8c6-8685-4205-b2f4-36195acfadaa" path="/var/lib/kubelet/pods/7b6ce8c6-8685-4205-b2f4-36195acfadaa/volumes" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.859711 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b","Type":"ContainerDied","Data":"d0e7754238fc27e029c3977aa70dfe65a538511fd5691cdaa817a039fb403edc"} Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.859789 4837 scope.go:117] "RemoveContainer" containerID="7c682076542a3610bbe65bd23558f10104bc0ccb17908bc4de955087dac16e7b" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.859734 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.891730 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.910646 4837 scope.go:117] "RemoveContainer" containerID="4c6ee7b9ec34581e58388a51448a539c6885341912eebcad0a4e9636b9d11b61" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.913606 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.939400 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.943164 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.949618 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.952588 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:06:23 crc kubenswrapper[4837]: I1001 09:06:23.952797 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.024459 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:24 crc kubenswrapper[4837]: W1001 09:06:24.030840 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ea0fb8_95b5_4e42_9872_5b209d41e33e.slice/crio-493385fa518bb14abc7f8f9f4b347050c271003f6ae00d67a5b0496434ec5118 WatchSource:0}: Error finding container 493385fa518bb14abc7f8f9f4b347050c271003f6ae00d67a5b0496434ec5118: Status 404 returned error can't find the container with id 493385fa518bb14abc7f8f9f4b347050c271003f6ae00d67a5b0496434ec5118 Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.060869 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.060931 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.061079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.061343 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.061583 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knlk2\" (UniqueName: \"kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.061715 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.061759 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.165834 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knlk2\" (UniqueName: \"kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166046 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166085 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166181 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166221 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166257 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.166377 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.194068 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.194485 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.195119 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.210120 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.213250 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.213542 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.213904 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knlk2\" (UniqueName: \"kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2\") pod \"glance-default-external-api-0\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.261528 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.840792 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:06:24 crc kubenswrapper[4837]: W1001 09:06:24.849044 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cef128b_5dd5_4e78_b9b8_4dfd86fbf31a.slice/crio-9e63e420d74a667f78d1d315086d49f3651e72c6de8bfad1de1f707f0196bc9d WatchSource:0}: Error finding container 9e63e420d74a667f78d1d315086d49f3651e72c6de8bfad1de1f707f0196bc9d: Status 404 returned error can't find the container with id 9e63e420d74a667f78d1d315086d49f3651e72c6de8bfad1de1f707f0196bc9d Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.881392 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerStarted","Data":"9e63e420d74a667f78d1d315086d49f3651e72c6de8bfad1de1f707f0196bc9d"} Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.886205 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerStarted","Data":"149de283f6d383b4a6f9334d505a3e6a4edd6e8aea2798cb6417bac072411ff6"} Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.886254 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerStarted","Data":"493385fa518bb14abc7f8f9f4b347050c271003f6ae00d67a5b0496434ec5118"} Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.891438 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerStarted","Data":"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442"} Oct 01 09:06:24 crc kubenswrapper[4837]: I1001 09:06:24.908451 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q2t4n" podStartSLOduration=4.099882804 podStartE2EDuration="10.908430265s" podCreationTimestamp="2025-10-01 09:06:14 +0000 UTC" firstStartedPulling="2025-10-01 09:06:16.701105204 +0000 UTC m=+7233.542712659" lastFinishedPulling="2025-10-01 09:06:23.509652665 +0000 UTC m=+7240.351260120" observedRunningTime="2025-10-01 09:06:24.906119858 +0000 UTC m=+7241.747727323" watchObservedRunningTime="2025-10-01 09:06:24.908430265 +0000 UTC m=+7241.750037720" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.042841 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.042883 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.300878 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.300923 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.304888 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.361061 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.826982 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b" path="/var/lib/kubelet/pods/d0913ad8-e8a5-4fd0-94c1-36a6f0b2190b/volumes" Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.907087 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerStarted","Data":"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d"} Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.909736 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerStarted","Data":"70c3598a3cf60fb7537aeba1afa09b81fffbe28781da4fefb255b244555e2158"} Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.909822 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-log" containerID="cri-o://149de283f6d383b4a6f9334d505a3e6a4edd6e8aea2798cb6417bac072411ff6" gracePeriod=30 Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.909866 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-httpd" containerID="cri-o://70c3598a3cf60fb7537aeba1afa09b81fffbe28781da4fefb255b244555e2158" gracePeriod=30 Oct 01 09:06:25 crc kubenswrapper[4837]: I1001 09:06:25.939881 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.939865631 podStartE2EDuration="3.939865631s" podCreationTimestamp="2025-10-01 09:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:25.93447271 +0000 UTC m=+7242.776080175" watchObservedRunningTime="2025-10-01 09:06:25.939865631 +0000 UTC m=+7242.781473086" Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.099523 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q2t4n" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="registry-server" probeResult="failure" output=< Oct 01 09:06:26 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:06:26 crc kubenswrapper[4837]: > Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.356899 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.418914 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.419208 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="dnsmasq-dns" containerID="cri-o://8678c653aa5843282f890228a0a4127405fb6217649ad0916d4d7720c4db631d" gracePeriod=10 Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.927069 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerStarted","Data":"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5"} Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.941603 4837 generic.go:334] "Generic (PLEG): container finished" podID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerID="8678c653aa5843282f890228a0a4127405fb6217649ad0916d4d7720c4db631d" exitCode=0 Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.941669 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" event={"ID":"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d","Type":"ContainerDied","Data":"8678c653aa5843282f890228a0a4127405fb6217649ad0916d4d7720c4db631d"} Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.941712 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" event={"ID":"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d","Type":"ContainerDied","Data":"e9994ddc852466630c5ad58e2ec1f88642f8f73322db1adc50e101917959703a"} Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.941723 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9994ddc852466630c5ad58e2ec1f88642f8f73322db1adc50e101917959703a" Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.951506 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.951488912 podStartE2EDuration="3.951488912s" podCreationTimestamp="2025-10-01 09:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:26.944010709 +0000 UTC m=+7243.785618164" watchObservedRunningTime="2025-10-01 09:06:26.951488912 +0000 UTC m=+7243.793096367" Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.954060 4837 generic.go:334] "Generic (PLEG): container finished" podID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerID="70c3598a3cf60fb7537aeba1afa09b81fffbe28781da4fefb255b244555e2158" exitCode=0 Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.954100 4837 generic.go:334] "Generic (PLEG): container finished" podID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerID="149de283f6d383b4a6f9334d505a3e6a4edd6e8aea2798cb6417bac072411ff6" exitCode=143 Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.955049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerDied","Data":"70c3598a3cf60fb7537aeba1afa09b81fffbe28781da4fefb255b244555e2158"} Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.955085 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerDied","Data":"149de283f6d383b4a6f9334d505a3e6a4edd6e8aea2798cb6417bac072411ff6"} Oct 01 09:06:26 crc kubenswrapper[4837]: I1001 09:06:26.986224 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:26.999491 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.043764 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb\") pod \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.043935 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.043969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5m4j\" (UniqueName: \"kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j\") pod \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044048 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb\") pod \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044102 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044244 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc\") pod \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044269 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7pln\" (UniqueName: \"kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044285 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044326 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044349 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config\") pod \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\" (UID: \"c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.044422 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts\") pod \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\" (UID: \"b0ea0fb8-95b5-4e42-9872-5b209d41e33e\") " Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.049790 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs" (OuterVolumeSpecName: "logs") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.062703 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.071908 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j" (OuterVolumeSpecName: "kube-api-access-t5m4j") pod "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" (UID: "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d"). InnerVolumeSpecName "kube-api-access-t5m4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.077859 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts" (OuterVolumeSpecName: "scripts") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.079941 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln" (OuterVolumeSpecName: "kube-api-access-r7pln") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "kube-api-access-r7pln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.086294 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.102912 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.125002 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" (UID: "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.127533 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" (UID: "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.130191 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" (UID: "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.133124 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.136684 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config" (OuterVolumeSpecName: "config") pod "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" (UID: "c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.140484 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data" (OuterVolumeSpecName: "config-data") pod "b0ea0fb8-95b5-4e42-9872-5b209d41e33e" (UID: "b0ea0fb8-95b5-4e42-9872-5b209d41e33e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.141206 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146443 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146470 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146480 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146490 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7pln\" (UniqueName: \"kubernetes.io/projected/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-kube-api-access-r7pln\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146499 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146511 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146522 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146532 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146542 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146553 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ea0fb8-95b5-4e42-9872-5b209d41e33e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146565 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5m4j\" (UniqueName: \"kubernetes.io/projected/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-kube-api-access-t5m4j\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.146575 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.816981 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:06:27 crc kubenswrapper[4837]: E1001 09:06:27.818082 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.969973 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.970041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b0ea0fb8-95b5-4e42-9872-5b209d41e33e","Type":"ContainerDied","Data":"493385fa518bb14abc7f8f9f4b347050c271003f6ae00d67a5b0496434ec5118"} Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.970119 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2d7sd" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="registry-server" containerID="cri-o://2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f" gracePeriod=2 Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.970185 4837 scope.go:117] "RemoveContainer" containerID="70c3598a3cf60fb7537aeba1afa09b81fffbe28781da4fefb255b244555e2158" Oct 01 09:06:27 crc kubenswrapper[4837]: I1001 09:06:27.970223 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b464cb87-qcrc7" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.031773 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.032924 4837 scope.go:117] "RemoveContainer" containerID="149de283f6d383b4a6f9334d505a3e6a4edd6e8aea2798cb6417bac072411ff6" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.043413 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66b464cb87-qcrc7"] Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.057990 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.070087 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087021 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:28 crc kubenswrapper[4837]: E1001 09:06:28.087428 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="init" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087459 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="init" Oct 01 09:06:28 crc kubenswrapper[4837]: E1001 09:06:28.087477 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-httpd" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087483 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-httpd" Oct 01 09:06:28 crc kubenswrapper[4837]: E1001 09:06:28.087493 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-log" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087499 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-log" Oct 01 09:06:28 crc kubenswrapper[4837]: E1001 09:06:28.087513 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="dnsmasq-dns" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087518 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="dnsmasq-dns" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087687 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-log" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087716 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" containerName="glance-httpd" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.087722 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" containerName="dnsmasq-dns" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.088741 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.093922 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.094030 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.128399 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.169834 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.169938 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdrq4\" (UniqueName: \"kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.169972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.170045 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.170098 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.170132 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.170152 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271410 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271503 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271539 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271601 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271653 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdrq4\" (UniqueName: \"kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.271674 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.272233 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.277308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.277628 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.278976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.280048 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.293078 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdrq4\" (UniqueName: \"kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.294313 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.467605 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.476506 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.575736 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities\") pod \"428efb5d-0b17-4645-86fa-d35446ab5d90\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.576349 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities" (OuterVolumeSpecName: "utilities") pod "428efb5d-0b17-4645-86fa-d35446ab5d90" (UID: "428efb5d-0b17-4645-86fa-d35446ab5d90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.576612 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content\") pod \"428efb5d-0b17-4645-86fa-d35446ab5d90\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.576720 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f58l\" (UniqueName: \"kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l\") pod \"428efb5d-0b17-4645-86fa-d35446ab5d90\" (UID: \"428efb5d-0b17-4645-86fa-d35446ab5d90\") " Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.577301 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.592072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l" (OuterVolumeSpecName: "kube-api-access-9f58l") pod "428efb5d-0b17-4645-86fa-d35446ab5d90" (UID: "428efb5d-0b17-4645-86fa-d35446ab5d90"). InnerVolumeSpecName "kube-api-access-9f58l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.622055 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "428efb5d-0b17-4645-86fa-d35446ab5d90" (UID: "428efb5d-0b17-4645-86fa-d35446ab5d90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.679403 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428efb5d-0b17-4645-86fa-d35446ab5d90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.679436 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f58l\" (UniqueName: \"kubernetes.io/projected/428efb5d-0b17-4645-86fa-d35446ab5d90-kube-api-access-9f58l\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.995055 4837 generic.go:334] "Generic (PLEG): container finished" podID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerID="2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f" exitCode=0 Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.995303 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2d7sd" Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.995338 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerDied","Data":"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f"} Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.996989 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2d7sd" event={"ID":"428efb5d-0b17-4645-86fa-d35446ab5d90","Type":"ContainerDied","Data":"701c37c59ad64b33c471758de80213b593d97fe1912ab50d8aaec6ee7d3511bd"} Oct 01 09:06:28 crc kubenswrapper[4837]: I1001 09:06:28.997032 4837 scope.go:117] "RemoveContainer" containerID="2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.035312 4837 scope.go:117] "RemoveContainer" containerID="e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.040011 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.049044 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2d7sd"] Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.070838 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.073061 4837 scope.go:117] "RemoveContainer" containerID="6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce" Oct 01 09:06:29 crc kubenswrapper[4837]: W1001 09:06:29.076620 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bdb56c8_a1bd_402b_89dd_4b0aacadf1a8.slice/crio-0f7dd090eb480a56ecf28044b3bc51d595d76e2c0cb5587f3508a306c8af7c58 WatchSource:0}: Error finding container 0f7dd090eb480a56ecf28044b3bc51d595d76e2c0cb5587f3508a306c8af7c58: Status 404 returned error can't find the container with id 0f7dd090eb480a56ecf28044b3bc51d595d76e2c0cb5587f3508a306c8af7c58 Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.221309 4837 scope.go:117] "RemoveContainer" containerID="2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f" Oct 01 09:06:29 crc kubenswrapper[4837]: E1001 09:06:29.222026 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f\": container with ID starting with 2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f not found: ID does not exist" containerID="2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.222154 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f"} err="failed to get container status \"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f\": rpc error: code = NotFound desc = could not find container \"2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f\": container with ID starting with 2e082ebcc416d4e2122280057da5f3d32a7e0b206caa59050fe2dada7e33941f not found: ID does not exist" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.222369 4837 scope.go:117] "RemoveContainer" containerID="e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a" Oct 01 09:06:29 crc kubenswrapper[4837]: E1001 09:06:29.223015 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a\": container with ID starting with e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a not found: ID does not exist" containerID="e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.223083 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a"} err="failed to get container status \"e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a\": rpc error: code = NotFound desc = could not find container \"e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a\": container with ID starting with e3aa1660a811edc9566ffb1d3f28f561c937257b64bc96029f47c2b29eac840a not found: ID does not exist" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.223116 4837 scope.go:117] "RemoveContainer" containerID="6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce" Oct 01 09:06:29 crc kubenswrapper[4837]: E1001 09:06:29.223596 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce\": container with ID starting with 6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce not found: ID does not exist" containerID="6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.223631 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce"} err="failed to get container status \"6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce\": rpc error: code = NotFound desc = could not find container \"6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce\": container with ID starting with 6aac01e6f152fa3c5ff802549e5bb653b230c5d87a57bc3370619cc2d0270bce not found: ID does not exist" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.826422 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" path="/var/lib/kubelet/pods/428efb5d-0b17-4645-86fa-d35446ab5d90/volumes" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.827552 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ea0fb8-95b5-4e42-9872-5b209d41e33e" path="/var/lib/kubelet/pods/b0ea0fb8-95b5-4e42-9872-5b209d41e33e/volumes" Oct 01 09:06:29 crc kubenswrapper[4837]: I1001 09:06:29.828138 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d" path="/var/lib/kubelet/pods/c317cb7c-4c0d-49d2-bae5-ad8a5fbdc92d/volumes" Oct 01 09:06:30 crc kubenswrapper[4837]: I1001 09:06:30.005907 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerStarted","Data":"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5"} Oct 01 09:06:30 crc kubenswrapper[4837]: I1001 09:06:30.006330 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerStarted","Data":"0f7dd090eb480a56ecf28044b3bc51d595d76e2c0cb5587f3508a306c8af7c58"} Oct 01 09:06:31 crc kubenswrapper[4837]: I1001 09:06:31.023404 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerStarted","Data":"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7"} Oct 01 09:06:31 crc kubenswrapper[4837]: I1001 09:06:31.064874 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.064840916 podStartE2EDuration="3.064840916s" podCreationTimestamp="2025-10-01 09:06:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:06:31.058630603 +0000 UTC m=+7247.900238118" watchObservedRunningTime="2025-10-01 09:06:31.064840916 +0000 UTC m=+7247.906448401" Oct 01 09:06:34 crc kubenswrapper[4837]: I1001 09:06:34.262663 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:06:34 crc kubenswrapper[4837]: I1001 09:06:34.264827 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:06:34 crc kubenswrapper[4837]: I1001 09:06:34.311975 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:06:34 crc kubenswrapper[4837]: I1001 09:06:34.314808 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.076620 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.076752 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.128019 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.207355 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.368449 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:35 crc kubenswrapper[4837]: I1001 09:06:35.371294 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:36 crc kubenswrapper[4837]: I1001 09:06:36.935499 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:06:36 crc kubenswrapper[4837]: I1001 09:06:36.939212 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.117133 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q2t4n" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="registry-server" containerID="cri-o://6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442" gracePeriod=2 Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.555831 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.675182 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8lvs\" (UniqueName: \"kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs\") pod \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.675386 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities\") pod \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.675446 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content\") pod \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\" (UID: \"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48\") " Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.676290 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities" (OuterVolumeSpecName: "utilities") pod "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" (UID: "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.676880 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.684142 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs" (OuterVolumeSpecName: "kube-api-access-r8lvs") pod "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" (UID: "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48"). InnerVolumeSpecName "kube-api-access-r8lvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.756340 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" (UID: "cefdf4bf-27a6-405d-b9e7-b718fc3dbb48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.771644 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.771970 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-znjtf" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="registry-server" containerID="cri-o://28e75f209fdcf6915703e965b75f6118853439623447694211142bbf4a202d0a" gracePeriod=2 Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.779089 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:37 crc kubenswrapper[4837]: I1001 09:06:37.779119 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8lvs\" (UniqueName: \"kubernetes.io/projected/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48-kube-api-access-r8lvs\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.107259 4837 generic.go:334] "Generic (PLEG): container finished" podID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerID="28e75f209fdcf6915703e965b75f6118853439623447694211142bbf4a202d0a" exitCode=0 Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.107337 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerDied","Data":"28e75f209fdcf6915703e965b75f6118853439623447694211142bbf4a202d0a"} Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.113425 4837 generic.go:334] "Generic (PLEG): container finished" podID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerID="6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442" exitCode=0 Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.113898 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t4n" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.114434 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerDied","Data":"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442"} Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.114460 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t4n" event={"ID":"cefdf4bf-27a6-405d-b9e7-b718fc3dbb48","Type":"ContainerDied","Data":"46283682e5b0571bd5aeb0ebe8087bdf508a457a7767b2237b11fa00b4b8ffe2"} Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.114496 4837 scope.go:117] "RemoveContainer" containerID="6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.139989 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.140710 4837 scope.go:117] "RemoveContainer" containerID="e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.151392 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q2t4n"] Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.163055 4837 scope.go:117] "RemoveContainer" containerID="9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.175393 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.193639 4837 scope.go:117] "RemoveContainer" containerID="6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442" Oct 01 09:06:38 crc kubenswrapper[4837]: E1001 09:06:38.194129 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442\": container with ID starting with 6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442 not found: ID does not exist" containerID="6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.194162 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442"} err="failed to get container status \"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442\": rpc error: code = NotFound desc = could not find container \"6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442\": container with ID starting with 6f9b913b0f3ed4cc56d99c4402da9fc632fe93f0f0d2928cffafbffe4dc55442 not found: ID does not exist" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.194189 4837 scope.go:117] "RemoveContainer" containerID="e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db" Oct 01 09:06:38 crc kubenswrapper[4837]: E1001 09:06:38.194536 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db\": container with ID starting with e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db not found: ID does not exist" containerID="e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.194559 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db"} err="failed to get container status \"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db\": rpc error: code = NotFound desc = could not find container \"e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db\": container with ID starting with e09696743bda4fe1392bb813596c51888537a03ffd740f706eda365def5558db not found: ID does not exist" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.194575 4837 scope.go:117] "RemoveContainer" containerID="9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d" Oct 01 09:06:38 crc kubenswrapper[4837]: E1001 09:06:38.194827 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d\": container with ID starting with 9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d not found: ID does not exist" containerID="9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.194849 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d"} err="failed to get container status \"9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d\": rpc error: code = NotFound desc = could not find container \"9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d\": container with ID starting with 9531f403efc7615603995d084cc6d056906bfa6dbb9cbbba849be1ffbb6b3f9d not found: ID does not exist" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.289196 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content\") pod \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.289431 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjfpp\" (UniqueName: \"kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp\") pod \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.289596 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities\") pod \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\" (UID: \"bbe9ae5b-243c-41c7-b832-501956dc2e2d\") " Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.290827 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities" (OuterVolumeSpecName: "utilities") pod "bbe9ae5b-243c-41c7-b832-501956dc2e2d" (UID: "bbe9ae5b-243c-41c7-b832-501956dc2e2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.298231 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp" (OuterVolumeSpecName: "kube-api-access-xjfpp") pod "bbe9ae5b-243c-41c7-b832-501956dc2e2d" (UID: "bbe9ae5b-243c-41c7-b832-501956dc2e2d"). InnerVolumeSpecName "kube-api-access-xjfpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.330574 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbe9ae5b-243c-41c7-b832-501956dc2e2d" (UID: "bbe9ae5b-243c-41c7-b832-501956dc2e2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.392278 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.392320 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjfpp\" (UniqueName: \"kubernetes.io/projected/bbe9ae5b-243c-41c7-b832-501956dc2e2d-kube-api-access-xjfpp\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.392335 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbe9ae5b-243c-41c7-b832-501956dc2e2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.468355 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.468453 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.520583 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:38 crc kubenswrapper[4837]: I1001 09:06:38.552313 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.131110 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znjtf" event={"ID":"bbe9ae5b-243c-41c7-b832-501956dc2e2d","Type":"ContainerDied","Data":"cef436345f85ac1e8a8d7c31e4d7feffcc98dc22a14b8a354b2b4b0a595b46bf"} Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.131627 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.131668 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.131172 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znjtf" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.131752 4837 scope.go:117] "RemoveContainer" containerID="28e75f209fdcf6915703e965b75f6118853439623447694211142bbf4a202d0a" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.176248 4837 scope.go:117] "RemoveContainer" containerID="63b0393d058faa43534c9838df4bc5172697b51db818348385271c07178ae2fd" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.198997 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.206239 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-znjtf"] Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.223347 4837 scope.go:117] "RemoveContainer" containerID="6cc321bee8f13b9f7a06548504ff31a32125034c0e7f22010712b6592bb31b9a" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.816742 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:06:39 crc kubenswrapper[4837]: E1001 09:06:39.817468 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.837103 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" path="/var/lib/kubelet/pods/bbe9ae5b-243c-41c7-b832-501956dc2e2d/volumes" Oct 01 09:06:39 crc kubenswrapper[4837]: I1001 09:06:39.839658 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" path="/var/lib/kubelet/pods/cefdf4bf-27a6-405d-b9e7-b718fc3dbb48/volumes" Oct 01 09:06:40 crc kubenswrapper[4837]: I1001 09:06:40.900418 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:40 crc kubenswrapper[4837]: I1001 09:06:40.933682 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.022050 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-6m6cr"] Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023409 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023441 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023505 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023526 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023541 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023555 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023583 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023603 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023641 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023654 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023685 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023782 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023801 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023814 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="extract-utilities" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023844 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023857 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: E1001 09:06:47.023876 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.023888 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="extract-content" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.024223 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe9ae5b-243c-41c7-b832-501956dc2e2d" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.024244 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="428efb5d-0b17-4645-86fa-d35446ab5d90" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.024281 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cefdf4bf-27a6-405d-b9e7-b718fc3dbb48" containerName="registry-server" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.025346 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.036404 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6m6cr"] Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.194748 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8t9q\" (UniqueName: \"kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q\") pod \"placement-db-create-6m6cr\" (UID: \"700833d9-e163-4dfa-b285-1493e045e28e\") " pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.297378 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8t9q\" (UniqueName: \"kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q\") pod \"placement-db-create-6m6cr\" (UID: \"700833d9-e163-4dfa-b285-1493e045e28e\") " pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.355826 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8t9q\" (UniqueName: \"kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q\") pod \"placement-db-create-6m6cr\" (UID: \"700833d9-e163-4dfa-b285-1493e045e28e\") " pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.654971 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:47 crc kubenswrapper[4837]: W1001 09:06:47.961985 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod700833d9_e163_4dfa_b285_1493e045e28e.slice/crio-df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b WatchSource:0}: Error finding container df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b: Status 404 returned error can't find the container with id df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b Oct 01 09:06:47 crc kubenswrapper[4837]: I1001 09:06:47.962067 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6m6cr"] Oct 01 09:06:48 crc kubenswrapper[4837]: I1001 09:06:48.234245 4837 generic.go:334] "Generic (PLEG): container finished" podID="700833d9-e163-4dfa-b285-1493e045e28e" containerID="9c04d6d846b070c5bfba0beccce8fa1b061aaf63d55e43edf65d61e957f58c90" exitCode=0 Oct 01 09:06:48 crc kubenswrapper[4837]: I1001 09:06:48.234334 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6m6cr" event={"ID":"700833d9-e163-4dfa-b285-1493e045e28e","Type":"ContainerDied","Data":"9c04d6d846b070c5bfba0beccce8fa1b061aaf63d55e43edf65d61e957f58c90"} Oct 01 09:06:48 crc kubenswrapper[4837]: I1001 09:06:48.234379 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6m6cr" event={"ID":"700833d9-e163-4dfa-b285-1493e045e28e","Type":"ContainerStarted","Data":"df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b"} Oct 01 09:06:49 crc kubenswrapper[4837]: I1001 09:06:49.682742 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:49 crc kubenswrapper[4837]: I1001 09:06:49.846892 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8t9q\" (UniqueName: \"kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q\") pod \"700833d9-e163-4dfa-b285-1493e045e28e\" (UID: \"700833d9-e163-4dfa-b285-1493e045e28e\") " Oct 01 09:06:49 crc kubenswrapper[4837]: I1001 09:06:49.857232 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q" (OuterVolumeSpecName: "kube-api-access-n8t9q") pod "700833d9-e163-4dfa-b285-1493e045e28e" (UID: "700833d9-e163-4dfa-b285-1493e045e28e"). InnerVolumeSpecName "kube-api-access-n8t9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:49 crc kubenswrapper[4837]: I1001 09:06:49.951740 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8t9q\" (UniqueName: \"kubernetes.io/projected/700833d9-e163-4dfa-b285-1493e045e28e-kube-api-access-n8t9q\") on node \"crc\" DevicePath \"\"" Oct 01 09:06:50 crc kubenswrapper[4837]: I1001 09:06:50.263203 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6m6cr" event={"ID":"700833d9-e163-4dfa-b285-1493e045e28e","Type":"ContainerDied","Data":"df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b"} Oct 01 09:06:50 crc kubenswrapper[4837]: I1001 09:06:50.263276 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df72697b83b8027c8cd600e29aa17c0ccefe4e81fade0818c8ea4bc09461541b" Oct 01 09:06:50 crc kubenswrapper[4837]: I1001 09:06:50.263370 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6m6cr" Oct 01 09:06:52 crc kubenswrapper[4837]: I1001 09:06:52.817790 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:06:52 crc kubenswrapper[4837]: E1001 09:06:52.819273 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.174325 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9a77-account-create-jsw66"] Oct 01 09:06:57 crc kubenswrapper[4837]: E1001 09:06:57.175304 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700833d9-e163-4dfa-b285-1493e045e28e" containerName="mariadb-database-create" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.175326 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="700833d9-e163-4dfa-b285-1493e045e28e" containerName="mariadb-database-create" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.175622 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="700833d9-e163-4dfa-b285-1493e045e28e" containerName="mariadb-database-create" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.176616 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.179296 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.199441 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9a77-account-create-jsw66"] Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.336673 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvsv8\" (UniqueName: \"kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8\") pod \"placement-9a77-account-create-jsw66\" (UID: \"512b846a-2ff8-44e3-b22f-f27dbe4d1571\") " pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.438350 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvsv8\" (UniqueName: \"kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8\") pod \"placement-9a77-account-create-jsw66\" (UID: \"512b846a-2ff8-44e3-b22f-f27dbe4d1571\") " pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.474783 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvsv8\" (UniqueName: \"kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8\") pod \"placement-9a77-account-create-jsw66\" (UID: \"512b846a-2ff8-44e3-b22f-f27dbe4d1571\") " pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:57 crc kubenswrapper[4837]: I1001 09:06:57.510084 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:58 crc kubenswrapper[4837]: I1001 09:06:58.010540 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9a77-account-create-jsw66"] Oct 01 09:06:58 crc kubenswrapper[4837]: I1001 09:06:58.356120 4837 generic.go:334] "Generic (PLEG): container finished" podID="512b846a-2ff8-44e3-b22f-f27dbe4d1571" containerID="f7b842d57e503b7e204b4b9518b076623531d2074b4518c54b8e81269b8d4112" exitCode=0 Oct 01 09:06:58 crc kubenswrapper[4837]: I1001 09:06:58.356269 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9a77-account-create-jsw66" event={"ID":"512b846a-2ff8-44e3-b22f-f27dbe4d1571","Type":"ContainerDied","Data":"f7b842d57e503b7e204b4b9518b076623531d2074b4518c54b8e81269b8d4112"} Oct 01 09:06:58 crc kubenswrapper[4837]: I1001 09:06:58.356838 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9a77-account-create-jsw66" event={"ID":"512b846a-2ff8-44e3-b22f-f27dbe4d1571","Type":"ContainerStarted","Data":"1c794ff6f2fcbeaddb45c1fbf51780ad699d9294ade14cea6330ebaed322b0a2"} Oct 01 09:06:59 crc kubenswrapper[4837]: I1001 09:06:59.767072 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:06:59 crc kubenswrapper[4837]: I1001 09:06:59.887147 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvsv8\" (UniqueName: \"kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8\") pod \"512b846a-2ff8-44e3-b22f-f27dbe4d1571\" (UID: \"512b846a-2ff8-44e3-b22f-f27dbe4d1571\") " Oct 01 09:06:59 crc kubenswrapper[4837]: I1001 09:06:59.896358 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8" (OuterVolumeSpecName: "kube-api-access-bvsv8") pod "512b846a-2ff8-44e3-b22f-f27dbe4d1571" (UID: "512b846a-2ff8-44e3-b22f-f27dbe4d1571"). InnerVolumeSpecName "kube-api-access-bvsv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:06:59 crc kubenswrapper[4837]: I1001 09:06:59.990513 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvsv8\" (UniqueName: \"kubernetes.io/projected/512b846a-2ff8-44e3-b22f-f27dbe4d1571-kube-api-access-bvsv8\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:00 crc kubenswrapper[4837]: I1001 09:07:00.389242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9a77-account-create-jsw66" event={"ID":"512b846a-2ff8-44e3-b22f-f27dbe4d1571","Type":"ContainerDied","Data":"1c794ff6f2fcbeaddb45c1fbf51780ad699d9294ade14cea6330ebaed322b0a2"} Oct 01 09:07:00 crc kubenswrapper[4837]: I1001 09:07:00.389299 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9a77-account-create-jsw66" Oct 01 09:07:00 crc kubenswrapper[4837]: I1001 09:07:00.389310 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c794ff6f2fcbeaddb45c1fbf51780ad699d9294ade14cea6330ebaed322b0a2" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.445852 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-n277b"] Oct 01 09:07:02 crc kubenswrapper[4837]: E1001 09:07:02.446543 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512b846a-2ff8-44e3-b22f-f27dbe4d1571" containerName="mariadb-account-create" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.446562 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="512b846a-2ff8-44e3-b22f-f27dbe4d1571" containerName="mariadb-account-create" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.447002 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="512b846a-2ff8-44e3-b22f-f27dbe4d1571" containerName="mariadb-account-create" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.447920 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.450073 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.450540 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.450768 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8dmm" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.464801 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n277b"] Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.484981 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.486567 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.505487 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.540438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42r2j\" (UniqueName: \"kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.540525 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.540559 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.540582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.540608 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641567 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641619 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gbjg\" (UniqueName: \"kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641656 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42r2j\" (UniqueName: \"kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641815 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641857 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641880 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641906 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641926 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641949 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.641972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.643201 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.653225 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.653459 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.653501 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.669220 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42r2j\" (UniqueName: \"kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j\") pod \"placement-db-sync-n277b\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743075 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gbjg\" (UniqueName: \"kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743390 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743425 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743457 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.743991 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.744308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.744760 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.744936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.763912 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gbjg\" (UniqueName: \"kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg\") pod \"dnsmasq-dns-7d58b9845-j5hrd\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.781467 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n277b" Oct 01 09:07:02 crc kubenswrapper[4837]: I1001 09:07:02.813205 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:03 crc kubenswrapper[4837]: I1001 09:07:03.277171 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n277b"] Oct 01 09:07:03 crc kubenswrapper[4837]: I1001 09:07:03.329733 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:07:03 crc kubenswrapper[4837]: I1001 09:07:03.418269 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n277b" event={"ID":"4b96552d-f2ac-43c9-9d2e-786da4133388","Type":"ContainerStarted","Data":"5df492cc2e74bc83d885c415725dd24b1fca63dc6260aa5a0c7f347f43fda0a1"} Oct 01 09:07:03 crc kubenswrapper[4837]: I1001 09:07:03.419483 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" event={"ID":"c2dc9d74-5044-4242-bf5e-8b5221d0b143","Type":"ContainerStarted","Data":"8c3675672bcd78347d64674a7af45b7e588122b42dc62d137724dca1be911526"} Oct 01 09:07:04 crc kubenswrapper[4837]: I1001 09:07:04.438607 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerID="9926a8ade6c8d86b1ceb312826fc2b62482a8c6032e0bce19f1b7e9e0c8cb390" exitCode=0 Oct 01 09:07:04 crc kubenswrapper[4837]: I1001 09:07:04.438934 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" event={"ID":"c2dc9d74-5044-4242-bf5e-8b5221d0b143","Type":"ContainerDied","Data":"9926a8ade6c8d86b1ceb312826fc2b62482a8c6032e0bce19f1b7e9e0c8cb390"} Oct 01 09:07:04 crc kubenswrapper[4837]: I1001 09:07:04.816620 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:07:04 crc kubenswrapper[4837]: E1001 09:07:04.817208 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:07:05 crc kubenswrapper[4837]: I1001 09:07:05.459511 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" event={"ID":"c2dc9d74-5044-4242-bf5e-8b5221d0b143","Type":"ContainerStarted","Data":"6e445beb75794c28ddc18870ed5e310001d0e1740b7ab1256fbf8d17b972c39f"} Oct 01 09:07:05 crc kubenswrapper[4837]: I1001 09:07:05.460772 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:05 crc kubenswrapper[4837]: I1001 09:07:05.486789 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" podStartSLOduration=3.486769774 podStartE2EDuration="3.486769774s" podCreationTimestamp="2025-10-01 09:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:07:05.480950251 +0000 UTC m=+7282.322557726" watchObservedRunningTime="2025-10-01 09:07:05.486769774 +0000 UTC m=+7282.328377239" Oct 01 09:07:07 crc kubenswrapper[4837]: I1001 09:07:07.487419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n277b" event={"ID":"4b96552d-f2ac-43c9-9d2e-786da4133388","Type":"ContainerStarted","Data":"ca4aed4b33ab5da98a7b94d1d093d97e7bfd7448ca44a7ba5ed34030ec3f3af8"} Oct 01 09:07:07 crc kubenswrapper[4837]: I1001 09:07:07.512194 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-n277b" podStartSLOduration=2.3560730579999998 podStartE2EDuration="5.512169117s" podCreationTimestamp="2025-10-01 09:07:02 +0000 UTC" firstStartedPulling="2025-10-01 09:07:03.278762504 +0000 UTC m=+7280.120369979" lastFinishedPulling="2025-10-01 09:07:06.434858573 +0000 UTC m=+7283.276466038" observedRunningTime="2025-10-01 09:07:07.506823486 +0000 UTC m=+7284.348430971" watchObservedRunningTime="2025-10-01 09:07:07.512169117 +0000 UTC m=+7284.353776612" Oct 01 09:07:08 crc kubenswrapper[4837]: I1001 09:07:08.502283 4837 generic.go:334] "Generic (PLEG): container finished" podID="4b96552d-f2ac-43c9-9d2e-786da4133388" containerID="ca4aed4b33ab5da98a7b94d1d093d97e7bfd7448ca44a7ba5ed34030ec3f3af8" exitCode=0 Oct 01 09:07:08 crc kubenswrapper[4837]: I1001 09:07:08.502352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n277b" event={"ID":"4b96552d-f2ac-43c9-9d2e-786da4133388","Type":"ContainerDied","Data":"ca4aed4b33ab5da98a7b94d1d093d97e7bfd7448ca44a7ba5ed34030ec3f3af8"} Oct 01 09:07:09 crc kubenswrapper[4837]: I1001 09:07:09.897786 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n277b" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002074 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42r2j\" (UniqueName: \"kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j\") pod \"4b96552d-f2ac-43c9-9d2e-786da4133388\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002266 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts\") pod \"4b96552d-f2ac-43c9-9d2e-786da4133388\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002313 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs\") pod \"4b96552d-f2ac-43c9-9d2e-786da4133388\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002343 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle\") pod \"4b96552d-f2ac-43c9-9d2e-786da4133388\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002471 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data\") pod \"4b96552d-f2ac-43c9-9d2e-786da4133388\" (UID: \"4b96552d-f2ac-43c9-9d2e-786da4133388\") " Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.002950 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs" (OuterVolumeSpecName: "logs") pod "4b96552d-f2ac-43c9-9d2e-786da4133388" (UID: "4b96552d-f2ac-43c9-9d2e-786da4133388"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.008800 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts" (OuterVolumeSpecName: "scripts") pod "4b96552d-f2ac-43c9-9d2e-786da4133388" (UID: "4b96552d-f2ac-43c9-9d2e-786da4133388"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.010703 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j" (OuterVolumeSpecName: "kube-api-access-42r2j") pod "4b96552d-f2ac-43c9-9d2e-786da4133388" (UID: "4b96552d-f2ac-43c9-9d2e-786da4133388"). InnerVolumeSpecName "kube-api-access-42r2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.047216 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data" (OuterVolumeSpecName: "config-data") pod "4b96552d-f2ac-43c9-9d2e-786da4133388" (UID: "4b96552d-f2ac-43c9-9d2e-786da4133388"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.054949 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b96552d-f2ac-43c9-9d2e-786da4133388" (UID: "4b96552d-f2ac-43c9-9d2e-786da4133388"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.104205 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.104239 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42r2j\" (UniqueName: \"kubernetes.io/projected/4b96552d-f2ac-43c9-9d2e-786da4133388-kube-api-access-42r2j\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.104251 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.104262 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b96552d-f2ac-43c9-9d2e-786da4133388-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.104270 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b96552d-f2ac-43c9-9d2e-786da4133388-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.525339 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n277b" event={"ID":"4b96552d-f2ac-43c9-9d2e-786da4133388","Type":"ContainerDied","Data":"5df492cc2e74bc83d885c415725dd24b1fca63dc6260aa5a0c7f347f43fda0a1"} Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.525391 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5df492cc2e74bc83d885c415725dd24b1fca63dc6260aa5a0c7f347f43fda0a1" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.525547 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n277b" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.637470 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-85d78dc59b-6dx5d"] Oct 01 09:07:10 crc kubenswrapper[4837]: E1001 09:07:10.638312 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b96552d-f2ac-43c9-9d2e-786da4133388" containerName="placement-db-sync" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.638425 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b96552d-f2ac-43c9-9d2e-786da4133388" containerName="placement-db-sync" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.638845 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b96552d-f2ac-43c9-9d2e-786da4133388" containerName="placement-db-sync" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.640552 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.645659 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.645972 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.646151 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.646332 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.646519 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8dmm" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.671631 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85d78dc59b-6dx5d"] Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.817893 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-combined-ca-bundle\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818328 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-config-data\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818372 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302856a7-e2da-4d70-b4f9-21b24570eeaa-logs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818522 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-scripts\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818548 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-public-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818623 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-482pg\" (UniqueName: \"kubernetes.io/projected/302856a7-e2da-4d70-b4f9-21b24570eeaa-kube-api-access-482pg\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.818687 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-internal-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920079 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-combined-ca-bundle\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920269 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-config-data\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302856a7-e2da-4d70-b4f9-21b24570eeaa-logs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920425 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-scripts\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920453 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-public-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920511 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-482pg\" (UniqueName: \"kubernetes.io/projected/302856a7-e2da-4d70-b4f9-21b24570eeaa-kube-api-access-482pg\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.920554 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-internal-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.921345 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302856a7-e2da-4d70-b4f9-21b24570eeaa-logs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.925685 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-combined-ca-bundle\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.925748 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-config-data\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.925725 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-scripts\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.927972 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-internal-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.928097 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/302856a7-e2da-4d70-b4f9-21b24570eeaa-public-tls-certs\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.947001 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-482pg\" (UniqueName: \"kubernetes.io/projected/302856a7-e2da-4d70-b4f9-21b24570eeaa-kube-api-access-482pg\") pod \"placement-85d78dc59b-6dx5d\" (UID: \"302856a7-e2da-4d70-b4f9-21b24570eeaa\") " pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:10 crc kubenswrapper[4837]: I1001 09:07:10.968393 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:11 crc kubenswrapper[4837]: I1001 09:07:11.558034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85d78dc59b-6dx5d"] Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.547888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85d78dc59b-6dx5d" event={"ID":"302856a7-e2da-4d70-b4f9-21b24570eeaa","Type":"ContainerStarted","Data":"640ea1770b251d04d546f3df39d119585d24d18ae4f17d8cb2010a2e7b62ad19"} Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.548394 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.548413 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85d78dc59b-6dx5d" event={"ID":"302856a7-e2da-4d70-b4f9-21b24570eeaa","Type":"ContainerStarted","Data":"dd5b61bea0be6695a62122def0219e94a396a0ba72fea85825f23630c57eb0d0"} Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.548428 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85d78dc59b-6dx5d" event={"ID":"302856a7-e2da-4d70-b4f9-21b24570eeaa","Type":"ContainerStarted","Data":"d1f2a0d7973c42860c37620ed6b623a30e702b091f830624481179bbcac3a583"} Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.548443 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.568618 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-85d78dc59b-6dx5d" podStartSLOduration=2.568600027 podStartE2EDuration="2.568600027s" podCreationTimestamp="2025-10-01 09:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:07:12.566435485 +0000 UTC m=+7289.408042950" watchObservedRunningTime="2025-10-01 09:07:12.568600027 +0000 UTC m=+7289.410207492" Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.814797 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.879775 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:07:12 crc kubenswrapper[4837]: I1001 09:07:12.880146 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="dnsmasq-dns" containerID="cri-o://61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18" gracePeriod=10 Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.332079 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.383629 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc\") pod \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.383880 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb\") pod \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.383940 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpj2t\" (UniqueName: \"kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t\") pod \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.384017 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config\") pod \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.384120 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb\") pod \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\" (UID: \"829d0b84-e85a-4eb8-a3e5-8bb3502e194f\") " Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.394542 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t" (OuterVolumeSpecName: "kube-api-access-xpj2t") pod "829d0b84-e85a-4eb8-a3e5-8bb3502e194f" (UID: "829d0b84-e85a-4eb8-a3e5-8bb3502e194f"). InnerVolumeSpecName "kube-api-access-xpj2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.457834 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "829d0b84-e85a-4eb8-a3e5-8bb3502e194f" (UID: "829d0b84-e85a-4eb8-a3e5-8bb3502e194f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.463296 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "829d0b84-e85a-4eb8-a3e5-8bb3502e194f" (UID: "829d0b84-e85a-4eb8-a3e5-8bb3502e194f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.468811 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "829d0b84-e85a-4eb8-a3e5-8bb3502e194f" (UID: "829d0b84-e85a-4eb8-a3e5-8bb3502e194f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.472859 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config" (OuterVolumeSpecName: "config") pod "829d0b84-e85a-4eb8-a3e5-8bb3502e194f" (UID: "829d0b84-e85a-4eb8-a3e5-8bb3502e194f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.490791 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.490851 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.490862 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.490872 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpj2t\" (UniqueName: \"kubernetes.io/projected/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-kube-api-access-xpj2t\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.490883 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/829d0b84-e85a-4eb8-a3e5-8bb3502e194f-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.567162 4837 generic.go:334] "Generic (PLEG): container finished" podID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerID="61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18" exitCode=0 Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.568074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" event={"ID":"829d0b84-e85a-4eb8-a3e5-8bb3502e194f","Type":"ContainerDied","Data":"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18"} Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.568109 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.568178 4837 scope.go:117] "RemoveContainer" containerID="61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.568159 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c9cf6db59-ts944" event={"ID":"829d0b84-e85a-4eb8-a3e5-8bb3502e194f","Type":"ContainerDied","Data":"f933b7c398ed4941be951e491759bca3e983c22a4ba4abfb040dd396988a396e"} Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.602884 4837 scope.go:117] "RemoveContainer" containerID="8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.615564 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.624751 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c9cf6db59-ts944"] Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.642654 4837 scope.go:117] "RemoveContainer" containerID="61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18" Oct 01 09:07:13 crc kubenswrapper[4837]: E1001 09:07:13.643206 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18\": container with ID starting with 61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18 not found: ID does not exist" containerID="61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.643260 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18"} err="failed to get container status \"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18\": rpc error: code = NotFound desc = could not find container \"61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18\": container with ID starting with 61c7fd74ca6bdf859079756ae34c0b8c6c0a8e4bad5a9c6d3a91c672b6154b18 not found: ID does not exist" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.643301 4837 scope.go:117] "RemoveContainer" containerID="8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e" Oct 01 09:07:13 crc kubenswrapper[4837]: E1001 09:07:13.643847 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e\": container with ID starting with 8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e not found: ID does not exist" containerID="8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.643883 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e"} err="failed to get container status \"8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e\": rpc error: code = NotFound desc = could not find container \"8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e\": container with ID starting with 8c3935447ac08e10a1ba2244e29432b6a805438abd0d0d19d0345abf5ecdfe9e not found: ID does not exist" Oct 01 09:07:13 crc kubenswrapper[4837]: I1001 09:07:13.832814 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" path="/var/lib/kubelet/pods/829d0b84-e85a-4eb8-a3e5-8bb3502e194f/volumes" Oct 01 09:07:17 crc kubenswrapper[4837]: I1001 09:07:17.817030 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:07:17 crc kubenswrapper[4837]: E1001 09:07:17.817895 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.006988 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:25 crc kubenswrapper[4837]: E1001 09:07:25.008088 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="init" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.008103 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="init" Oct 01 09:07:25 crc kubenswrapper[4837]: E1001 09:07:25.008142 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="dnsmasq-dns" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.008149 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="dnsmasq-dns" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.008308 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="829d0b84-e85a-4eb8-a3e5-8bb3502e194f" containerName="dnsmasq-dns" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.009606 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.024085 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.136749 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8pt5\" (UniqueName: \"kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.136826 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.136880 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.239204 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8pt5\" (UniqueName: \"kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.239278 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.239341 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.239905 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.239952 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.266417 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8pt5\" (UniqueName: \"kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5\") pod \"redhat-marketplace-4qckz\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.331326 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:25 crc kubenswrapper[4837]: I1001 09:07:25.775823 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:25 crc kubenswrapper[4837]: W1001 09:07:25.783124 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62c79591_46dc_40eb_8a6e_e9f6824d8ae4.slice/crio-3c15dfba2be98c9ef118f5686ccb4cad0e215ebedeff8393e72d426585d5a730 WatchSource:0}: Error finding container 3c15dfba2be98c9ef118f5686ccb4cad0e215ebedeff8393e72d426585d5a730: Status 404 returned error can't find the container with id 3c15dfba2be98c9ef118f5686ccb4cad0e215ebedeff8393e72d426585d5a730 Oct 01 09:07:26 crc kubenswrapper[4837]: I1001 09:07:26.730680 4837 generic.go:334] "Generic (PLEG): container finished" podID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerID="bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0" exitCode=0 Oct 01 09:07:26 crc kubenswrapper[4837]: I1001 09:07:26.730756 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerDied","Data":"bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0"} Oct 01 09:07:26 crc kubenswrapper[4837]: I1001 09:07:26.730788 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerStarted","Data":"3c15dfba2be98c9ef118f5686ccb4cad0e215ebedeff8393e72d426585d5a730"} Oct 01 09:07:28 crc kubenswrapper[4837]: I1001 09:07:28.753414 4837 generic.go:334] "Generic (PLEG): container finished" podID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerID="235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62" exitCode=0 Oct 01 09:07:28 crc kubenswrapper[4837]: I1001 09:07:28.753542 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerDied","Data":"235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62"} Oct 01 09:07:29 crc kubenswrapper[4837]: I1001 09:07:29.767204 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerStarted","Data":"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7"} Oct 01 09:07:29 crc kubenswrapper[4837]: I1001 09:07:29.809543 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4qckz" podStartSLOduration=3.064905236 podStartE2EDuration="5.809527917s" podCreationTimestamp="2025-10-01 09:07:24 +0000 UTC" firstStartedPulling="2025-10-01 09:07:26.733052424 +0000 UTC m=+7303.574659919" lastFinishedPulling="2025-10-01 09:07:29.477675105 +0000 UTC m=+7306.319282600" observedRunningTime="2025-10-01 09:07:29.806616496 +0000 UTC m=+7306.648223961" watchObservedRunningTime="2025-10-01 09:07:29.809527917 +0000 UTC m=+7306.651135372" Oct 01 09:07:29 crc kubenswrapper[4837]: I1001 09:07:29.815801 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:07:29 crc kubenswrapper[4837]: E1001 09:07:29.816114 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:07:35 crc kubenswrapper[4837]: I1001 09:07:35.332257 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:35 crc kubenswrapper[4837]: I1001 09:07:35.332782 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:35 crc kubenswrapper[4837]: I1001 09:07:35.380263 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:35 crc kubenswrapper[4837]: I1001 09:07:35.889545 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:35 crc kubenswrapper[4837]: I1001 09:07:35.944637 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:37 crc kubenswrapper[4837]: I1001 09:07:37.848741 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4qckz" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="registry-server" containerID="cri-o://4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7" gracePeriod=2 Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.362434 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.516867 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8pt5\" (UniqueName: \"kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5\") pod \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.516969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities\") pod \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.517204 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content\") pod \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\" (UID: \"62c79591-46dc-40eb-8a6e-e9f6824d8ae4\") " Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.518756 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities" (OuterVolumeSpecName: "utilities") pod "62c79591-46dc-40eb-8a6e-e9f6824d8ae4" (UID: "62c79591-46dc-40eb-8a6e-e9f6824d8ae4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.526241 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5" (OuterVolumeSpecName: "kube-api-access-z8pt5") pod "62c79591-46dc-40eb-8a6e-e9f6824d8ae4" (UID: "62c79591-46dc-40eb-8a6e-e9f6824d8ae4"). InnerVolumeSpecName "kube-api-access-z8pt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.546302 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62c79591-46dc-40eb-8a6e-e9f6824d8ae4" (UID: "62c79591-46dc-40eb-8a6e-e9f6824d8ae4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.619524 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.619556 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8pt5\" (UniqueName: \"kubernetes.io/projected/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-kube-api-access-z8pt5\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.619568 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c79591-46dc-40eb-8a6e-e9f6824d8ae4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.863140 4837 generic.go:334] "Generic (PLEG): container finished" podID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerID="4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7" exitCode=0 Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.863202 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerDied","Data":"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7"} Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.863256 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4qckz" event={"ID":"62c79591-46dc-40eb-8a6e-e9f6824d8ae4","Type":"ContainerDied","Data":"3c15dfba2be98c9ef118f5686ccb4cad0e215ebedeff8393e72d426585d5a730"} Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.863283 4837 scope.go:117] "RemoveContainer" containerID="4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.863214 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4qckz" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.897713 4837 scope.go:117] "RemoveContainer" containerID="235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.923401 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.934308 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4qckz"] Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.944688 4837 scope.go:117] "RemoveContainer" containerID="bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.981787 4837 scope.go:117] "RemoveContainer" containerID="4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7" Oct 01 09:07:38 crc kubenswrapper[4837]: E1001 09:07:38.982621 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7\": container with ID starting with 4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7 not found: ID does not exist" containerID="4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.982724 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7"} err="failed to get container status \"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7\": rpc error: code = NotFound desc = could not find container \"4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7\": container with ID starting with 4b3869297b1401988f68ed95f5d3557949c1010ee7790490452809e58ca5aeb7 not found: ID does not exist" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.982767 4837 scope.go:117] "RemoveContainer" containerID="235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62" Oct 01 09:07:38 crc kubenswrapper[4837]: E1001 09:07:38.983315 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62\": container with ID starting with 235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62 not found: ID does not exist" containerID="235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.983365 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62"} err="failed to get container status \"235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62\": rpc error: code = NotFound desc = could not find container \"235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62\": container with ID starting with 235bbf1b778cc08b28d0b9eb665cf2204ee3b1af035586747f7de8f7335bfa62 not found: ID does not exist" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.983401 4837 scope.go:117] "RemoveContainer" containerID="bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0" Oct 01 09:07:38 crc kubenswrapper[4837]: E1001 09:07:38.983914 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0\": container with ID starting with bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0 not found: ID does not exist" containerID="bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0" Oct 01 09:07:38 crc kubenswrapper[4837]: I1001 09:07:38.983948 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0"} err="failed to get container status \"bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0\": rpc error: code = NotFound desc = could not find container \"bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0\": container with ID starting with bddd53f9b07a380ca0a1826d7c45d1f4bfa102ce0b2d3b3ad122bf566a76dad0 not found: ID does not exist" Oct 01 09:07:39 crc kubenswrapper[4837]: I1001 09:07:39.837506 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" path="/var/lib/kubelet/pods/62c79591-46dc-40eb-8a6e-e9f6824d8ae4/volumes" Oct 01 09:07:41 crc kubenswrapper[4837]: I1001 09:07:41.923936 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:41 crc kubenswrapper[4837]: I1001 09:07:41.933667 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85d78dc59b-6dx5d" Oct 01 09:07:43 crc kubenswrapper[4837]: I1001 09:07:43.826099 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:07:43 crc kubenswrapper[4837]: E1001 09:07:43.826932 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:07:57 crc kubenswrapper[4837]: I1001 09:07:57.815681 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:07:58 crc kubenswrapper[4837]: I1001 09:07:58.089212 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810"} Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.192128 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8n975"] Oct 01 09:08:02 crc kubenswrapper[4837]: E1001 09:08:02.192965 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="extract-content" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.192981 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="extract-content" Oct 01 09:08:02 crc kubenswrapper[4837]: E1001 09:08:02.193016 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="extract-utilities" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.193024 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="extract-utilities" Oct 01 09:08:02 crc kubenswrapper[4837]: E1001 09:08:02.193041 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="registry-server" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.193048 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="registry-server" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.193248 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c79591-46dc-40eb-8a6e-e9f6824d8ae4" containerName="registry-server" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.193981 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.208516 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8n975"] Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.229963 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxq94\" (UniqueName: \"kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94\") pod \"nova-api-db-create-8n975\" (UID: \"e7403bed-3c82-49c3-8d57-75470652e3db\") " pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.298710 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-qjgjc"] Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.299832 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.309822 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qjgjc"] Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.332302 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxq94\" (UniqueName: \"kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94\") pod \"nova-api-db-create-8n975\" (UID: \"e7403bed-3c82-49c3-8d57-75470652e3db\") " pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.332379 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjlrg\" (UniqueName: \"kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg\") pod \"nova-cell0-db-create-qjgjc\" (UID: \"c1095f7c-8690-4423-a608-0aeca0c1871c\") " pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.360566 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxq94\" (UniqueName: \"kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94\") pod \"nova-api-db-create-8n975\" (UID: \"e7403bed-3c82-49c3-8d57-75470652e3db\") " pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.434784 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjlrg\" (UniqueName: \"kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg\") pod \"nova-cell0-db-create-qjgjc\" (UID: \"c1095f7c-8690-4423-a608-0aeca0c1871c\") " pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.475099 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjlrg\" (UniqueName: \"kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg\") pod \"nova-cell0-db-create-qjgjc\" (UID: \"c1095f7c-8690-4423-a608-0aeca0c1871c\") " pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.506575 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2wmlh"] Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.508819 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.515846 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.517449 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2wmlh"] Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.537399 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hb9\" (UniqueName: \"kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9\") pod \"nova-cell1-db-create-2wmlh\" (UID: \"e0e6452d-fee6-43ca-9845-43f6f30c2a29\") " pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.616683 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.642297 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hb9\" (UniqueName: \"kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9\") pod \"nova-cell1-db-create-2wmlh\" (UID: \"e0e6452d-fee6-43ca-9845-43f6f30c2a29\") " pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.663895 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hb9\" (UniqueName: \"kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9\") pod \"nova-cell1-db-create-2wmlh\" (UID: \"e0e6452d-fee6-43ca-9845-43f6f30c2a29\") " pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:02 crc kubenswrapper[4837]: I1001 09:08:02.944428 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:03 crc kubenswrapper[4837]: I1001 09:08:03.074781 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8n975"] Oct 01 09:08:03 crc kubenswrapper[4837]: I1001 09:08:03.143743 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8n975" event={"ID":"e7403bed-3c82-49c3-8d57-75470652e3db","Type":"ContainerStarted","Data":"6d62df43e201f27c8d6696c2a172ecca4c5cf92a3b77d9f8032e57467f9added"} Oct 01 09:08:03 crc kubenswrapper[4837]: I1001 09:08:03.209175 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qjgjc"] Oct 01 09:08:03 crc kubenswrapper[4837]: W1001 09:08:03.212995 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1095f7c_8690_4423_a608_0aeca0c1871c.slice/crio-80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315 WatchSource:0}: Error finding container 80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315: Status 404 returned error can't find the container with id 80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315 Oct 01 09:08:03 crc kubenswrapper[4837]: W1001 09:08:03.219420 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0e6452d_fee6_43ca_9845_43f6f30c2a29.slice/crio-8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928 WatchSource:0}: Error finding container 8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928: Status 404 returned error can't find the container with id 8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928 Oct 01 09:08:03 crc kubenswrapper[4837]: I1001 09:08:03.223017 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2wmlh"] Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.168916 4837 generic.go:334] "Generic (PLEG): container finished" podID="e0e6452d-fee6-43ca-9845-43f6f30c2a29" containerID="ae0488ee5ab012d506b74c065262a89f049462c1051efc738e936b5f1ba33a79" exitCode=0 Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.169109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2wmlh" event={"ID":"e0e6452d-fee6-43ca-9845-43f6f30c2a29","Type":"ContainerDied","Data":"ae0488ee5ab012d506b74c065262a89f049462c1051efc738e936b5f1ba33a79"} Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.169799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2wmlh" event={"ID":"e0e6452d-fee6-43ca-9845-43f6f30c2a29","Type":"ContainerStarted","Data":"8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928"} Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.174619 4837 generic.go:334] "Generic (PLEG): container finished" podID="c1095f7c-8690-4423-a608-0aeca0c1871c" containerID="c7c70daef7196a6c174366bf118012c84eb95952bb75c36be847e684a72cae9c" exitCode=0 Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.174741 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qjgjc" event={"ID":"c1095f7c-8690-4423-a608-0aeca0c1871c","Type":"ContainerDied","Data":"c7c70daef7196a6c174366bf118012c84eb95952bb75c36be847e684a72cae9c"} Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.175231 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qjgjc" event={"ID":"c1095f7c-8690-4423-a608-0aeca0c1871c","Type":"ContainerStarted","Data":"80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315"} Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.178570 4837 generic.go:334] "Generic (PLEG): container finished" podID="e7403bed-3c82-49c3-8d57-75470652e3db" containerID="cd96e17d423b561dc071a5ced945d6261327d8ca80de9dde1e9d81b38add842e" exitCode=0 Oct 01 09:08:04 crc kubenswrapper[4837]: I1001 09:08:04.178643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8n975" event={"ID":"e7403bed-3c82-49c3-8d57-75470652e3db","Type":"ContainerDied","Data":"cd96e17d423b561dc071a5ced945d6261327d8ca80de9dde1e9d81b38add842e"} Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.647651 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.652613 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.657801 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.693307 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjlrg\" (UniqueName: \"kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg\") pod \"c1095f7c-8690-4423-a608-0aeca0c1871c\" (UID: \"c1095f7c-8690-4423-a608-0aeca0c1871c\") " Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.693912 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxq94\" (UniqueName: \"kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94\") pod \"e7403bed-3c82-49c3-8d57-75470652e3db\" (UID: \"e7403bed-3c82-49c3-8d57-75470652e3db\") " Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.694256 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hb9\" (UniqueName: \"kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9\") pod \"e0e6452d-fee6-43ca-9845-43f6f30c2a29\" (UID: \"e0e6452d-fee6-43ca-9845-43f6f30c2a29\") " Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.702382 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94" (OuterVolumeSpecName: "kube-api-access-cxq94") pod "e7403bed-3c82-49c3-8d57-75470652e3db" (UID: "e7403bed-3c82-49c3-8d57-75470652e3db"). InnerVolumeSpecName "kube-api-access-cxq94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.702449 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg" (OuterVolumeSpecName: "kube-api-access-bjlrg") pod "c1095f7c-8690-4423-a608-0aeca0c1871c" (UID: "c1095f7c-8690-4423-a608-0aeca0c1871c"). InnerVolumeSpecName "kube-api-access-bjlrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.702867 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9" (OuterVolumeSpecName: "kube-api-access-68hb9") pod "e0e6452d-fee6-43ca-9845-43f6f30c2a29" (UID: "e0e6452d-fee6-43ca-9845-43f6f30c2a29"). InnerVolumeSpecName "kube-api-access-68hb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.797269 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hb9\" (UniqueName: \"kubernetes.io/projected/e0e6452d-fee6-43ca-9845-43f6f30c2a29-kube-api-access-68hb9\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.797310 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjlrg\" (UniqueName: \"kubernetes.io/projected/c1095f7c-8690-4423-a608-0aeca0c1871c-kube-api-access-bjlrg\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:05 crc kubenswrapper[4837]: I1001 09:08:05.797321 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxq94\" (UniqueName: \"kubernetes.io/projected/e7403bed-3c82-49c3-8d57-75470652e3db-kube-api-access-cxq94\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.205185 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qjgjc" event={"ID":"c1095f7c-8690-4423-a608-0aeca0c1871c","Type":"ContainerDied","Data":"80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315"} Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.205477 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80813fa2173f1353beb34fa5da8bcfa03d77628141316bba66baaeb95c086315" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.205191 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qjgjc" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.207948 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8n975" event={"ID":"e7403bed-3c82-49c3-8d57-75470652e3db","Type":"ContainerDied","Data":"6d62df43e201f27c8d6696c2a172ecca4c5cf92a3b77d9f8032e57467f9added"} Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.208038 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d62df43e201f27c8d6696c2a172ecca4c5cf92a3b77d9f8032e57467f9added" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.207972 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8n975" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.211335 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2wmlh" event={"ID":"e0e6452d-fee6-43ca-9845-43f6f30c2a29","Type":"ContainerDied","Data":"8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928"} Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.211378 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8138f924dd13f7469f67da3e8a459e186202a9895c44d095e65c0b697adba928" Oct 01 09:08:06 crc kubenswrapper[4837]: I1001 09:08:06.211397 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2wmlh" Oct 01 09:08:09 crc kubenswrapper[4837]: I1001 09:08:09.720994 4837 scope.go:117] "RemoveContainer" containerID="6dece9d37f8a1e31b9b030470eed6ea3af9c2829f0a36cd0812ff7104d3ae375" Oct 01 09:08:09 crc kubenswrapper[4837]: I1001 09:08:09.786016 4837 scope.go:117] "RemoveContainer" containerID="caeae328ad87f649439b55b74f29c599c5af547745f919fcda9612dd5da5c1ab" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.444671 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3a18-account-create-qfkdm"] Oct 01 09:08:12 crc kubenswrapper[4837]: E1001 09:08:12.445619 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1095f7c-8690-4423-a608-0aeca0c1871c" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445633 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1095f7c-8690-4423-a608-0aeca0c1871c" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: E1001 09:08:12.445652 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7403bed-3c82-49c3-8d57-75470652e3db" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445659 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7403bed-3c82-49c3-8d57-75470652e3db" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: E1001 09:08:12.445670 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e6452d-fee6-43ca-9845-43f6f30c2a29" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445676 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e6452d-fee6-43ca-9845-43f6f30c2a29" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445895 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1095f7c-8690-4423-a608-0aeca0c1871c" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445907 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e6452d-fee6-43ca-9845-43f6f30c2a29" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.445927 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7403bed-3c82-49c3-8d57-75470652e3db" containerName="mariadb-database-create" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.446498 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.448796 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.458324 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3a18-account-create-qfkdm"] Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.636948 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2ea8-account-create-p2hkr"] Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.638539 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.641438 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.641751 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thx6l\" (UniqueName: \"kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l\") pod \"nova-api-3a18-account-create-qfkdm\" (UID: \"9dd8a385-1c4d-4878-853d-3140eaf1fbd8\") " pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.655015 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2ea8-account-create-p2hkr"] Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.743919 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thx6l\" (UniqueName: \"kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l\") pod \"nova-api-3a18-account-create-qfkdm\" (UID: \"9dd8a385-1c4d-4878-853d-3140eaf1fbd8\") " pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.744113 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw76l\" (UniqueName: \"kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l\") pod \"nova-cell0-2ea8-account-create-p2hkr\" (UID: \"036173d9-f63f-4124-9278-594a6f412eb6\") " pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.774556 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thx6l\" (UniqueName: \"kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l\") pod \"nova-api-3a18-account-create-qfkdm\" (UID: \"9dd8a385-1c4d-4878-853d-3140eaf1fbd8\") " pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.776469 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.845452 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw76l\" (UniqueName: \"kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l\") pod \"nova-cell0-2ea8-account-create-p2hkr\" (UID: \"036173d9-f63f-4124-9278-594a6f412eb6\") " pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.857003 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cefb-account-create-6mgpb"] Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.858440 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.861191 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.873021 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw76l\" (UniqueName: \"kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l\") pod \"nova-cell0-2ea8-account-create-p2hkr\" (UID: \"036173d9-f63f-4124-9278-594a6f412eb6\") " pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.877803 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cefb-account-create-6mgpb"] Oct 01 09:08:12 crc kubenswrapper[4837]: I1001 09:08:12.976355 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.051787 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g5l26"] Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.052838 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kk54\" (UniqueName: \"kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54\") pod \"nova-cell1-cefb-account-create-6mgpb\" (UID: \"c2fad0f2-8515-4a39-96b4-adf4faf14545\") " pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.057132 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g5l26"] Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.156858 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kk54\" (UniqueName: \"kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54\") pod \"nova-cell1-cefb-account-create-6mgpb\" (UID: \"c2fad0f2-8515-4a39-96b4-adf4faf14545\") " pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.174430 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kk54\" (UniqueName: \"kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54\") pod \"nova-cell1-cefb-account-create-6mgpb\" (UID: \"c2fad0f2-8515-4a39-96b4-adf4faf14545\") " pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.250232 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.289520 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3a18-account-create-qfkdm"] Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.482588 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2ea8-account-create-p2hkr"] Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.528864 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cefb-account-create-6mgpb"] Oct 01 09:08:13 crc kubenswrapper[4837]: W1001 09:08:13.533382 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2fad0f2_8515_4a39_96b4_adf4faf14545.slice/crio-5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8 WatchSource:0}: Error finding container 5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8: Status 404 returned error can't find the container with id 5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8 Oct 01 09:08:13 crc kubenswrapper[4837]: I1001 09:08:13.829756 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71a25257-e61a-493c-ae6b-42a29d97a6cd" path="/var/lib/kubelet/pods/71a25257-e61a-493c-ae6b-42a29d97a6cd/volumes" Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.293036 4837 generic.go:334] "Generic (PLEG): container finished" podID="036173d9-f63f-4124-9278-594a6f412eb6" containerID="1d2102928b6c3afb9e9f1d360c6b17f3730feabe302ad315b3cf53004b32ba8f" exitCode=0 Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.293095 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" event={"ID":"036173d9-f63f-4124-9278-594a6f412eb6","Type":"ContainerDied","Data":"1d2102928b6c3afb9e9f1d360c6b17f3730feabe302ad315b3cf53004b32ba8f"} Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.293156 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" event={"ID":"036173d9-f63f-4124-9278-594a6f412eb6","Type":"ContainerStarted","Data":"d36eae46ebce0a38a638c43a9db0792599126614b42ca66e7abaab14382c3f8c"} Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.295842 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2fad0f2-8515-4a39-96b4-adf4faf14545" containerID="b2a2a1c1bcc52ffef6d39b999104801f7a2c7edc1123eb6d4388d0ee634f4564" exitCode=0 Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.295901 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cefb-account-create-6mgpb" event={"ID":"c2fad0f2-8515-4a39-96b4-adf4faf14545","Type":"ContainerDied","Data":"b2a2a1c1bcc52ffef6d39b999104801f7a2c7edc1123eb6d4388d0ee634f4564"} Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.295948 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cefb-account-create-6mgpb" event={"ID":"c2fad0f2-8515-4a39-96b4-adf4faf14545","Type":"ContainerStarted","Data":"5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8"} Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.298728 4837 generic.go:334] "Generic (PLEG): container finished" podID="9dd8a385-1c4d-4878-853d-3140eaf1fbd8" containerID="575d3a2180b5688693aec0ce3cc809d3baa4495b5243b96de56e531277ef64b1" exitCode=0 Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.298774 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a18-account-create-qfkdm" event={"ID":"9dd8a385-1c4d-4878-853d-3140eaf1fbd8","Type":"ContainerDied","Data":"575d3a2180b5688693aec0ce3cc809d3baa4495b5243b96de56e531277ef64b1"} Oct 01 09:08:14 crc kubenswrapper[4837]: I1001 09:08:14.298804 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a18-account-create-qfkdm" event={"ID":"9dd8a385-1c4d-4878-853d-3140eaf1fbd8","Type":"ContainerStarted","Data":"dd30523ebe48794e38efedd70dac97ea18fac699575b6491599c93c2e50da7f8"} Oct 01 09:08:15 crc kubenswrapper[4837]: I1001 09:08:15.834629 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.023182 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw76l\" (UniqueName: \"kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l\") pod \"036173d9-f63f-4124-9278-594a6f412eb6\" (UID: \"036173d9-f63f-4124-9278-594a6f412eb6\") " Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.024429 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.031886 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l" (OuterVolumeSpecName: "kube-api-access-zw76l") pod "036173d9-f63f-4124-9278-594a6f412eb6" (UID: "036173d9-f63f-4124-9278-594a6f412eb6"). InnerVolumeSpecName "kube-api-access-zw76l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.034078 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.124716 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kk54\" (UniqueName: \"kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54\") pod \"c2fad0f2-8515-4a39-96b4-adf4faf14545\" (UID: \"c2fad0f2-8515-4a39-96b4-adf4faf14545\") " Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.125057 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thx6l\" (UniqueName: \"kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l\") pod \"9dd8a385-1c4d-4878-853d-3140eaf1fbd8\" (UID: \"9dd8a385-1c4d-4878-853d-3140eaf1fbd8\") " Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.125362 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw76l\" (UniqueName: \"kubernetes.io/projected/036173d9-f63f-4124-9278-594a6f412eb6-kube-api-access-zw76l\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.127481 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54" (OuterVolumeSpecName: "kube-api-access-2kk54") pod "c2fad0f2-8515-4a39-96b4-adf4faf14545" (UID: "c2fad0f2-8515-4a39-96b4-adf4faf14545"). InnerVolumeSpecName "kube-api-access-2kk54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.137880 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l" (OuterVolumeSpecName: "kube-api-access-thx6l") pod "9dd8a385-1c4d-4878-853d-3140eaf1fbd8" (UID: "9dd8a385-1c4d-4878-853d-3140eaf1fbd8"). InnerVolumeSpecName "kube-api-access-thx6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.227246 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kk54\" (UniqueName: \"kubernetes.io/projected/c2fad0f2-8515-4a39-96b4-adf4faf14545-kube-api-access-2kk54\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.227284 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thx6l\" (UniqueName: \"kubernetes.io/projected/9dd8a385-1c4d-4878-853d-3140eaf1fbd8-kube-api-access-thx6l\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.318735 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3a18-account-create-qfkdm" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.318731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3a18-account-create-qfkdm" event={"ID":"9dd8a385-1c4d-4878-853d-3140eaf1fbd8","Type":"ContainerDied","Data":"dd30523ebe48794e38efedd70dac97ea18fac699575b6491599c93c2e50da7f8"} Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.318786 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd30523ebe48794e38efedd70dac97ea18fac699575b6491599c93c2e50da7f8" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.320298 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" event={"ID":"036173d9-f63f-4124-9278-594a6f412eb6","Type":"ContainerDied","Data":"d36eae46ebce0a38a638c43a9db0792599126614b42ca66e7abaab14382c3f8c"} Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.320359 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d36eae46ebce0a38a638c43a9db0792599126614b42ca66e7abaab14382c3f8c" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.320330 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2ea8-account-create-p2hkr" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.321819 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cefb-account-create-6mgpb" event={"ID":"c2fad0f2-8515-4a39-96b4-adf4faf14545","Type":"ContainerDied","Data":"5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8"} Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.321840 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aab6ef85f7644099070d70b8ac000f112dad33b3de933d6f4d04a5a6ce8a3f8" Oct 01 09:08:16 crc kubenswrapper[4837]: I1001 09:08:16.321888 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cefb-account-create-6mgpb" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859075 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qlgcs"] Oct 01 09:08:17 crc kubenswrapper[4837]: E1001 09:08:17.859642 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd8a385-1c4d-4878-853d-3140eaf1fbd8" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859654 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd8a385-1c4d-4878-853d-3140eaf1fbd8" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: E1001 09:08:17.859675 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036173d9-f63f-4124-9278-594a6f412eb6" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859681 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="036173d9-f63f-4124-9278-594a6f412eb6" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: E1001 09:08:17.859712 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2fad0f2-8515-4a39-96b4-adf4faf14545" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859718 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2fad0f2-8515-4a39-96b4-adf4faf14545" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859882 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2fad0f2-8515-4a39-96b4-adf4faf14545" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859897 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd8a385-1c4d-4878-853d-3140eaf1fbd8" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.859909 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="036173d9-f63f-4124-9278-594a6f412eb6" containerName="mariadb-account-create" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.860450 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.863834 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.864250 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wbm2h" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.864534 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.879033 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qlgcs"] Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.957180 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxtsw\" (UniqueName: \"kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.957303 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.957383 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:17 crc kubenswrapper[4837]: I1001 09:08:17.957402 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.058995 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.059096 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.059113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.059155 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxtsw\" (UniqueName: \"kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.064245 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.064407 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.070799 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.085976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxtsw\" (UniqueName: \"kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw\") pod \"nova-cell0-conductor-db-sync-qlgcs\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.179494 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:18 crc kubenswrapper[4837]: I1001 09:08:18.709674 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qlgcs"] Oct 01 09:08:18 crc kubenswrapper[4837]: W1001 09:08:18.721852 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a51e260_c980_4c8d_94d3_cd1f8e9c6f66.slice/crio-f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426 WatchSource:0}: Error finding container f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426: Status 404 returned error can't find the container with id f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426 Oct 01 09:08:19 crc kubenswrapper[4837]: I1001 09:08:19.354348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" event={"ID":"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66","Type":"ContainerStarted","Data":"f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426"} Oct 01 09:08:23 crc kubenswrapper[4837]: I1001 09:08:23.033500 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4fc3-account-create-4k5wb"] Oct 01 09:08:23 crc kubenswrapper[4837]: I1001 09:08:23.043800 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4fc3-account-create-4k5wb"] Oct 01 09:08:23 crc kubenswrapper[4837]: I1001 09:08:23.828119 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8347ab54-35b3-41a2-a132-83654b5ad74c" path="/var/lib/kubelet/pods/8347ab54-35b3-41a2-a132-83654b5ad74c/volumes" Oct 01 09:08:28 crc kubenswrapper[4837]: I1001 09:08:28.479835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" event={"ID":"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66","Type":"ContainerStarted","Data":"62791ee7dccb440098f077245e147c6f69d5319ed6da30e8620e5bb40529d030"} Oct 01 09:08:28 crc kubenswrapper[4837]: I1001 09:08:28.523437 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" podStartSLOduration=2.856791747 podStartE2EDuration="11.523422721s" podCreationTimestamp="2025-10-01 09:08:17 +0000 UTC" firstStartedPulling="2025-10-01 09:08:18.723643352 +0000 UTC m=+7355.565250807" lastFinishedPulling="2025-10-01 09:08:27.390274326 +0000 UTC m=+7364.231881781" observedRunningTime="2025-10-01 09:08:28.502547238 +0000 UTC m=+7365.344154693" watchObservedRunningTime="2025-10-01 09:08:28.523422721 +0000 UTC m=+7365.365030176" Oct 01 09:08:32 crc kubenswrapper[4837]: I1001 09:08:32.550804 4837 generic.go:334] "Generic (PLEG): container finished" podID="1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" containerID="62791ee7dccb440098f077245e147c6f69d5319ed6da30e8620e5bb40529d030" exitCode=0 Oct 01 09:08:32 crc kubenswrapper[4837]: I1001 09:08:32.550890 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" event={"ID":"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66","Type":"ContainerDied","Data":"62791ee7dccb440098f077245e147c6f69d5319ed6da30e8620e5bb40529d030"} Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.015018 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.126622 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts\") pod \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.126760 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data\") pod \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.126813 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxtsw\" (UniqueName: \"kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw\") pod \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.126901 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle\") pod \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\" (UID: \"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66\") " Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.134992 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts" (OuterVolumeSpecName: "scripts") pod "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" (UID: "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.136305 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw" (OuterVolumeSpecName: "kube-api-access-bxtsw") pod "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" (UID: "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66"). InnerVolumeSpecName "kube-api-access-bxtsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.162081 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data" (OuterVolumeSpecName: "config-data") pod "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" (UID: "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.162984 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" (UID: "1a51e260-c980-4c8d-94d3-cd1f8e9c6f66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.229371 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.229401 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.229413 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxtsw\" (UniqueName: \"kubernetes.io/projected/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-kube-api-access-bxtsw\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.229422 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.589277 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" event={"ID":"1a51e260-c980-4c8d-94d3-cd1f8e9c6f66","Type":"ContainerDied","Data":"f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426"} Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.589784 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1ecfb6b1d8d35ee72f3b6e5d669b195bf82d25f3d9069d291ac78c4ffe81426" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.589398 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-qlgcs" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.692526 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:08:34 crc kubenswrapper[4837]: E1001 09:08:34.693176 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" containerName="nova-cell0-conductor-db-sync" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.693215 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" containerName="nova-cell0-conductor-db-sync" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.693569 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" containerName="nova-cell0-conductor-db-sync" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.694560 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.697404 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wbm2h" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.697617 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.706969 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.842389 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.842517 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.842779 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st5jn\" (UniqueName: \"kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.946355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st5jn\" (UniqueName: \"kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.949557 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.950017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.956746 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.957755 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:34 crc kubenswrapper[4837]: I1001 09:08:34.965306 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st5jn\" (UniqueName: \"kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn\") pod \"nova-cell0-conductor-0\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.023925 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.057091 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dfl72"] Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.065045 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dfl72"] Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.541910 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:08:35 crc kubenswrapper[4837]: W1001 09:08:35.546053 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94df4b1a_2b87_4889_9713_fcc96cbfd579.slice/crio-4a705e1ab5bce2bd3d21fa8c5850415034ddeefc00447576f9f1338e2226b501 WatchSource:0}: Error finding container 4a705e1ab5bce2bd3d21fa8c5850415034ddeefc00447576f9f1338e2226b501: Status 404 returned error can't find the container with id 4a705e1ab5bce2bd3d21fa8c5850415034ddeefc00447576f9f1338e2226b501 Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.606263 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94df4b1a-2b87-4889-9713-fcc96cbfd579","Type":"ContainerStarted","Data":"4a705e1ab5bce2bd3d21fa8c5850415034ddeefc00447576f9f1338e2226b501"} Oct 01 09:08:35 crc kubenswrapper[4837]: I1001 09:08:35.824914 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0d2fd75-023a-4c79-beb1-e1a9393ba32e" path="/var/lib/kubelet/pods/b0d2fd75-023a-4c79-beb1-e1a9393ba32e/volumes" Oct 01 09:08:36 crc kubenswrapper[4837]: I1001 09:08:36.625551 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94df4b1a-2b87-4889-9713-fcc96cbfd579","Type":"ContainerStarted","Data":"3f219857e39bd485a4069aec9a14316e135d4dc2cd3edc38b0d094d07f572d5c"} Oct 01 09:08:36 crc kubenswrapper[4837]: I1001 09:08:36.625765 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:36 crc kubenswrapper[4837]: I1001 09:08:36.667008 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.666944739 podStartE2EDuration="2.666944739s" podCreationTimestamp="2025-10-01 09:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:36.653307102 +0000 UTC m=+7373.494914597" watchObservedRunningTime="2025-10-01 09:08:36.666944739 +0000 UTC m=+7373.508552274" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.077316 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.623329 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ktt7n"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.628366 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.630885 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ktt7n"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.631974 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.641489 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.674433 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.674533 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.674568 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.674637 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npt7s\" (UniqueName: \"kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.776366 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.776456 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.776489 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.776528 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npt7s\" (UniqueName: \"kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.786745 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.788155 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.790144 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.796995 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.809655 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.810119 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.815637 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.824428 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npt7s\" (UniqueName: \"kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s\") pod \"nova-cell0-cell-mapping-ktt7n\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.853384 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.865981 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.869933 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.876115 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.928455 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.929538 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.935190 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.942341 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.960039 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlq5r\" (UniqueName: \"kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981476 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981558 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981658 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981680 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:40 crc kubenswrapper[4837]: I1001 09:08:40.981734 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jctz\" (UniqueName: \"kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.031208 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.033148 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.035875 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.054015 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083133 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083204 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083250 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jctz\" (UniqueName: \"kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083276 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlq5r\" (UniqueName: \"kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083292 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083315 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083364 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083382 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.083434 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qd7d\" (UniqueName: \"kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.085245 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.092327 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.095230 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.102897 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.114469 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.124804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jctz\" (UniqueName: \"kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz\") pod \"nova-scheduler-0\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.130799 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.132755 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.134199 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlq5r\" (UniqueName: \"kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r\") pod \"nova-api-0\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.155591 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.185286 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.185617 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.185839 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.185965 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.186005 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qd7d\" (UniqueName: \"kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.186043 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkc9r\" (UniqueName: \"kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.186457 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.192452 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.192451 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.207308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qd7d\" (UniqueName: \"kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.207888 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.222887 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.262125 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289074 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkc9r\" (UniqueName: \"kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289209 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289234 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289332 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289394 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289420 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289463 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrdcr\" (UniqueName: \"kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.289514 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.291311 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.294638 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.296403 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.306237 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkc9r\" (UniqueName: \"kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r\") pod \"nova-metadata-0\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.385224 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.391709 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.391746 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.391762 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.391816 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.391844 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrdcr\" (UniqueName: \"kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.392615 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.394908 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.394986 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.395124 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.411165 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrdcr\" (UniqueName: \"kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr\") pod \"dnsmasq-dns-9d649f7dc-rn6j5\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.452540 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.548212 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ktt7n"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.705503 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ktt7n" event={"ID":"33918084-d6ec-433c-9fbe-8f1386fb347e","Type":"ContainerStarted","Data":"b337068ea6f92662254e301783f10c6f51ee25ffa0267f6084bf2edd43f07824"} Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.739388 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.772108 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pj4qg"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.774499 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.779482 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.779767 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.802755 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pj4qg"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.854945 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.877611 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.904566 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.904628 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.904650 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdcjb\" (UniqueName: \"kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.904892 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:41 crc kubenswrapper[4837]: I1001 09:08:41.966622 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:41 crc kubenswrapper[4837]: W1001 09:08:41.974914 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod921d0029_ecc2_4bda_8a79_5c7035c6c3d2.slice/crio-8f4cbf607522c9e17e31588b2b5bf29203fc75f4028083c2963ac153cfc7bb7e WatchSource:0}: Error finding container 8f4cbf607522c9e17e31588b2b5bf29203fc75f4028083c2963ac153cfc7bb7e: Status 404 returned error can't find the container with id 8f4cbf607522c9e17e31588b2b5bf29203fc75f4028083c2963ac153cfc7bb7e Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.007221 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.007384 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.007448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.007473 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdcjb\" (UniqueName: \"kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.017400 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.017528 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.017723 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.022903 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdcjb\" (UniqueName: \"kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb\") pod \"nova-cell1-conductor-db-sync-pj4qg\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.072527 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:08:42 crc kubenswrapper[4837]: W1001 09:08:42.097862 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadfd5762_5f62_4d1d_8e76_baa2f6ff7a91.slice/crio-395ea95b0e016e9bff142e8260e50644dcbc01bb9a58f7f022c8987425eb2ba5 WatchSource:0}: Error finding container 395ea95b0e016e9bff142e8260e50644dcbc01bb9a58f7f022c8987425eb2ba5: Status 404 returned error can't find the container with id 395ea95b0e016e9bff142e8260e50644dcbc01bb9a58f7f022c8987425eb2ba5 Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.149653 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.740748 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02d99b3a-47b2-4e39-b144-7659c300cc25","Type":"ContainerStarted","Data":"a57414b9b080d78233a94fc3b1d1a84b47507ec67ab9408b44b603206dd26569"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.747397 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6592600a-c30d-42ed-854a-22f489159c0e","Type":"ContainerStarted","Data":"3f430aaf6a1f27e692cdeeb4ba7dd3e36edffb953a9951cc9fff97459ec42496"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.748789 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerStarted","Data":"a88f70f8f05876ac908b5eff91e28cba052a2c86584cc4051ea28756795cba8c"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.750445 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ktt7n" event={"ID":"33918084-d6ec-433c-9fbe-8f1386fb347e","Type":"ContainerStarted","Data":"9e2d383f5619fa58b7348c2c88ef6434a049bfb1e28579cb4884eb0279ddae32"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.754836 4837 generic.go:334] "Generic (PLEG): container finished" podID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerID="4475635f91138bb3eb48b2e0c8530826a5717df6515d4e6a95d3a0de71fb8fb1" exitCode=0 Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.755142 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" event={"ID":"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91","Type":"ContainerDied","Data":"4475635f91138bb3eb48b2e0c8530826a5717df6515d4e6a95d3a0de71fb8fb1"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.755177 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" event={"ID":"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91","Type":"ContainerStarted","Data":"395ea95b0e016e9bff142e8260e50644dcbc01bb9a58f7f022c8987425eb2ba5"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.756908 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pj4qg"] Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.769866 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerStarted","Data":"8f4cbf607522c9e17e31588b2b5bf29203fc75f4028083c2963ac153cfc7bb7e"} Oct 01 09:08:42 crc kubenswrapper[4837]: I1001 09:08:42.775590 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ktt7n" podStartSLOduration=2.775566809 podStartE2EDuration="2.775566809s" podCreationTimestamp="2025-10-01 09:08:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:42.770381952 +0000 UTC m=+7379.611989407" watchObservedRunningTime="2025-10-01 09:08:42.775566809 +0000 UTC m=+7379.617174264" Oct 01 09:08:44 crc kubenswrapper[4837]: I1001 09:08:44.805486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" event={"ID":"cafbdbfe-2005-402e-a731-20f4afb49879","Type":"ContainerStarted","Data":"7b84e63f029563bc09930990a51d04d4ffef05e15e529ace168c9b9a0db1d703"} Oct 01 09:08:44 crc kubenswrapper[4837]: I1001 09:08:44.996749 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.007558 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.834361 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.834707 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerStarted","Data":"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.834731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerStarted","Data":"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.834744 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" event={"ID":"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91","Type":"ContainerStarted","Data":"6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.837481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerStarted","Data":"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.837528 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerStarted","Data":"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.837654 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-metadata" containerID="cri-o://47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" gracePeriod=30 Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.837644 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-log" containerID="cri-o://e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" gracePeriod=30 Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.844483 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02d99b3a-47b2-4e39-b144-7659c300cc25","Type":"ContainerStarted","Data":"eca879cf191100d8cde22e388f93e7afd846fe85c9cbddef53a6d3eb9fa32589"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.849552 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.999684663 podStartE2EDuration="5.849539882s" podCreationTimestamp="2025-10-01 09:08:40 +0000 UTC" firstStartedPulling="2025-10-01 09:08:41.759399633 +0000 UTC m=+7378.601007088" lastFinishedPulling="2025-10-01 09:08:44.609254852 +0000 UTC m=+7381.450862307" observedRunningTime="2025-10-01 09:08:45.844428555 +0000 UTC m=+7382.686036010" watchObservedRunningTime="2025-10-01 09:08:45.849539882 +0000 UTC m=+7382.691147337" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.851481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" event={"ID":"cafbdbfe-2005-402e-a731-20f4afb49879","Type":"ContainerStarted","Data":"1345222af4a5403f43eb2a1848d36f26b54d28ca266554c5443a182034f59ec2"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.866552 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6592600a-c30d-42ed-854a-22f489159c0e","Type":"ContainerStarted","Data":"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f"} Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.887254 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" podStartSLOduration=4.887229606 podStartE2EDuration="4.887229606s" podCreationTimestamp="2025-10-01 09:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:45.881297929 +0000 UTC m=+7382.722905384" watchObservedRunningTime="2025-10-01 09:08:45.887229606 +0000 UTC m=+7382.728837091" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.901547 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.211682105 podStartE2EDuration="4.90152941s" podCreationTimestamp="2025-10-01 09:08:41 +0000 UTC" firstStartedPulling="2025-10-01 09:08:41.977049055 +0000 UTC m=+7378.818656510" lastFinishedPulling="2025-10-01 09:08:44.66689636 +0000 UTC m=+7381.508503815" observedRunningTime="2025-10-01 09:08:45.897783617 +0000 UTC m=+7382.739391072" watchObservedRunningTime="2025-10-01 09:08:45.90152941 +0000 UTC m=+7382.743136865" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.924525 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" podStartSLOduration=4.924508699 podStartE2EDuration="4.924508699s" podCreationTimestamp="2025-10-01 09:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:45.923440433 +0000 UTC m=+7382.765047918" watchObservedRunningTime="2025-10-01 09:08:45.924508699 +0000 UTC m=+7382.766116154" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.954946 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.217958162 podStartE2EDuration="5.954679437s" podCreationTimestamp="2025-10-01 09:08:40 +0000 UTC" firstStartedPulling="2025-10-01 09:08:41.89528133 +0000 UTC m=+7378.736888785" lastFinishedPulling="2025-10-01 09:08:44.632002605 +0000 UTC m=+7381.473610060" observedRunningTime="2025-10-01 09:08:45.939711666 +0000 UTC m=+7382.781319121" watchObservedRunningTime="2025-10-01 09:08:45.954679437 +0000 UTC m=+7382.796286932" Oct 01 09:08:45 crc kubenswrapper[4837]: I1001 09:08:45.968977 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.190370996 podStartE2EDuration="5.96895073s" podCreationTimestamp="2025-10-01 09:08:40 +0000 UTC" firstStartedPulling="2025-10-01 09:08:41.855285728 +0000 UTC m=+7378.696893183" lastFinishedPulling="2025-10-01 09:08:44.633865462 +0000 UTC m=+7381.475472917" observedRunningTime="2025-10-01 09:08:45.963119416 +0000 UTC m=+7382.804726871" watchObservedRunningTime="2025-10-01 09:08:45.96895073 +0000 UTC m=+7382.810558195" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.223602 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.263113 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.385961 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.386134 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.463661 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.638377 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle\") pod \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.638507 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkc9r\" (UniqueName: \"kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r\") pod \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.638603 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs\") pod \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.638710 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data\") pod \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\" (UID: \"921d0029-ecc2-4bda-8a79-5c7035c6c3d2\") " Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.641086 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs" (OuterVolumeSpecName: "logs") pod "921d0029-ecc2-4bda-8a79-5c7035c6c3d2" (UID: "921d0029-ecc2-4bda-8a79-5c7035c6c3d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.658867 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r" (OuterVolumeSpecName: "kube-api-access-kkc9r") pod "921d0029-ecc2-4bda-8a79-5c7035c6c3d2" (UID: "921d0029-ecc2-4bda-8a79-5c7035c6c3d2"). InnerVolumeSpecName "kube-api-access-kkc9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.672713 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "921d0029-ecc2-4bda-8a79-5c7035c6c3d2" (UID: "921d0029-ecc2-4bda-8a79-5c7035c6c3d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.682333 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data" (OuterVolumeSpecName: "config-data") pod "921d0029-ecc2-4bda-8a79-5c7035c6c3d2" (UID: "921d0029-ecc2-4bda-8a79-5c7035c6c3d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.742084 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkc9r\" (UniqueName: \"kubernetes.io/projected/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-kube-api-access-kkc9r\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.742133 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.742149 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.742164 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/921d0029-ecc2-4bda-8a79-5c7035c6c3d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880632 4837 generic.go:334] "Generic (PLEG): container finished" podID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerID="47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" exitCode=0 Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880671 4837 generic.go:334] "Generic (PLEG): container finished" podID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerID="e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" exitCode=143 Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880740 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerDied","Data":"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe"} Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880777 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerDied","Data":"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb"} Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880792 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"921d0029-ecc2-4bda-8a79-5c7035c6c3d2","Type":"ContainerDied","Data":"8f4cbf607522c9e17e31588b2b5bf29203fc75f4028083c2963ac153cfc7bb7e"} Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880811 4837 scope.go:117] "RemoveContainer" containerID="47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.880979 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.896013 4837 generic.go:334] "Generic (PLEG): container finished" podID="33918084-d6ec-433c-9fbe-8f1386fb347e" containerID="9e2d383f5619fa58b7348c2c88ef6434a049bfb1e28579cb4884eb0279ddae32" exitCode=0 Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.896188 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ktt7n" event={"ID":"33918084-d6ec-433c-9fbe-8f1386fb347e","Type":"ContainerDied","Data":"9e2d383f5619fa58b7348c2c88ef6434a049bfb1e28579cb4884eb0279ddae32"} Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.896836 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6592600a-c30d-42ed-854a-22f489159c0e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f" gracePeriod=30 Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.924714 4837 scope.go:117] "RemoveContainer" containerID="e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.957284 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.972924 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.978642 4837 scope.go:117] "RemoveContainer" containerID="47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" Oct 01 09:08:46 crc kubenswrapper[4837]: E1001 09:08:46.979176 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe\": container with ID starting with 47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe not found: ID does not exist" containerID="47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.979227 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe"} err="failed to get container status \"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe\": rpc error: code = NotFound desc = could not find container \"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe\": container with ID starting with 47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe not found: ID does not exist" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.979259 4837 scope.go:117] "RemoveContainer" containerID="e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" Oct 01 09:08:46 crc kubenswrapper[4837]: E1001 09:08:46.979745 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb\": container with ID starting with e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb not found: ID does not exist" containerID="e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.979784 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb"} err="failed to get container status \"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb\": rpc error: code = NotFound desc = could not find container \"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb\": container with ID starting with e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb not found: ID does not exist" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.979804 4837 scope.go:117] "RemoveContainer" containerID="47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.980439 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe"} err="failed to get container status \"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe\": rpc error: code = NotFound desc = could not find container \"47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe\": container with ID starting with 47274f5d0ed108238209615a185b66abfeb4947ddabed90b35e580d259a5cfbe not found: ID does not exist" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.980487 4837 scope.go:117] "RemoveContainer" containerID="e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.980867 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb"} err="failed to get container status \"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb\": rpc error: code = NotFound desc = could not find container \"e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb\": container with ID starting with e401482868363c62d31fdb111e2c0dd671a26faade724316c96f9918ed9786fb not found: ID does not exist" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.983141 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:46 crc kubenswrapper[4837]: E1001 09:08:46.983606 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-log" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.983627 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-log" Oct 01 09:08:46 crc kubenswrapper[4837]: E1001 09:08:46.983637 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-metadata" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.983644 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-metadata" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.983888 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-metadata" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.983902 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" containerName="nova-metadata-log" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.984945 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:46 crc kubenswrapper[4837]: I1001 09:08:46.999682 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.004393 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.004643 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.149233 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.149366 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.149520 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.149548 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.149583 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfj7\" (UniqueName: \"kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.252645 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.253242 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.253863 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.254057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfj7\" (UniqueName: \"kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.254213 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.254480 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.259338 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.262522 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.262542 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.274267 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfj7\" (UniqueName: \"kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7\") pod \"nova-metadata-0\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.319419 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.825037 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="921d0029-ecc2-4bda-8a79-5c7035c6c3d2" path="/var/lib/kubelet/pods/921d0029-ecc2-4bda-8a79-5c7035c6c3d2/volumes" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.831633 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.899511 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:47 crc kubenswrapper[4837]: W1001 09:08:47.902427 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00fb0081_d5ce_4392_bdd2_cb0642cfb7b9.slice/crio-f76fa665ad2d0e617e3039f7442d9f9d6b63011ebac299c8d8f38b5d3f73fdfe WatchSource:0}: Error finding container f76fa665ad2d0e617e3039f7442d9f9d6b63011ebac299c8d8f38b5d3f73fdfe: Status 404 returned error can't find the container with id f76fa665ad2d0e617e3039f7442d9f9d6b63011ebac299c8d8f38b5d3f73fdfe Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.910641 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" event={"ID":"cafbdbfe-2005-402e-a731-20f4afb49879","Type":"ContainerDied","Data":"1345222af4a5403f43eb2a1848d36f26b54d28ca266554c5443a182034f59ec2"} Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.910554 4837 generic.go:334] "Generic (PLEG): container finished" podID="cafbdbfe-2005-402e-a731-20f4afb49879" containerID="1345222af4a5403f43eb2a1848d36f26b54d28ca266554c5443a182034f59ec2" exitCode=0 Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.912632 4837 generic.go:334] "Generic (PLEG): container finished" podID="6592600a-c30d-42ed-854a-22f489159c0e" containerID="7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f" exitCode=0 Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.912715 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.912714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6592600a-c30d-42ed-854a-22f489159c0e","Type":"ContainerDied","Data":"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f"} Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.912852 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6592600a-c30d-42ed-854a-22f489159c0e","Type":"ContainerDied","Data":"3f430aaf6a1f27e692cdeeb4ba7dd3e36edffb953a9951cc9fff97459ec42496"} Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.912879 4837 scope.go:117] "RemoveContainer" containerID="7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.968658 4837 scope.go:117] "RemoveContainer" containerID="7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f" Oct 01 09:08:47 crc kubenswrapper[4837]: E1001 09:08:47.969520 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f\": container with ID starting with 7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f not found: ID does not exist" containerID="7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.969597 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f"} err="failed to get container status \"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f\": rpc error: code = NotFound desc = could not find container \"7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f\": container with ID starting with 7c8a7a9abd742cef79860d5f0ddfc62a706f25f8b1a89746626c219b24d8004f not found: ID does not exist" Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.971220 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data\") pod \"6592600a-c30d-42ed-854a-22f489159c0e\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.971434 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle\") pod \"6592600a-c30d-42ed-854a-22f489159c0e\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.971517 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qd7d\" (UniqueName: \"kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d\") pod \"6592600a-c30d-42ed-854a-22f489159c0e\" (UID: \"6592600a-c30d-42ed-854a-22f489159c0e\") " Oct 01 09:08:47 crc kubenswrapper[4837]: I1001 09:08:47.979307 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d" (OuterVolumeSpecName: "kube-api-access-8qd7d") pod "6592600a-c30d-42ed-854a-22f489159c0e" (UID: "6592600a-c30d-42ed-854a-22f489159c0e"). InnerVolumeSpecName "kube-api-access-8qd7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.007822 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data" (OuterVolumeSpecName: "config-data") pod "6592600a-c30d-42ed-854a-22f489159c0e" (UID: "6592600a-c30d-42ed-854a-22f489159c0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.009832 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6592600a-c30d-42ed-854a-22f489159c0e" (UID: "6592600a-c30d-42ed-854a-22f489159c0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.076810 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.077081 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6592600a-c30d-42ed-854a-22f489159c0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.077094 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qd7d\" (UniqueName: \"kubernetes.io/projected/6592600a-c30d-42ed-854a-22f489159c0e-kube-api-access-8qd7d\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.249094 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.260947 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.265928 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.273853 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:48 crc kubenswrapper[4837]: E1001 09:08:48.274372 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33918084-d6ec-433c-9fbe-8f1386fb347e" containerName="nova-manage" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.274395 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="33918084-d6ec-433c-9fbe-8f1386fb347e" containerName="nova-manage" Oct 01 09:08:48 crc kubenswrapper[4837]: E1001 09:08:48.274414 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6592600a-c30d-42ed-854a-22f489159c0e" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.274420 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6592600a-c30d-42ed-854a-22f489159c0e" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.274651 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6592600a-c30d-42ed-854a-22f489159c0e" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.274672 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="33918084-d6ec-433c-9fbe-8f1386fb347e" containerName="nova-manage" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.276897 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279529 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle\") pod \"33918084-d6ec-433c-9fbe-8f1386fb347e\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279576 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts\") pod \"33918084-d6ec-433c-9fbe-8f1386fb347e\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279628 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npt7s\" (UniqueName: \"kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s\") pod \"33918084-d6ec-433c-9fbe-8f1386fb347e\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279671 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data\") pod \"33918084-d6ec-433c-9fbe-8f1386fb347e\" (UID: \"33918084-d6ec-433c-9fbe-8f1386fb347e\") " Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279874 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279967 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.279996 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx56n\" (UniqueName: \"kubernetes.io/projected/ba4d4a69-79d0-4f8b-83db-319fd0aca696-kube-api-access-xx56n\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.280028 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.280046 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.282221 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.283671 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.284068 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.285825 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s" (OuterVolumeSpecName: "kube-api-access-npt7s") pod "33918084-d6ec-433c-9fbe-8f1386fb347e" (UID: "33918084-d6ec-433c-9fbe-8f1386fb347e"). InnerVolumeSpecName "kube-api-access-npt7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.289718 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts" (OuterVolumeSpecName: "scripts") pod "33918084-d6ec-433c-9fbe-8f1386fb347e" (UID: "33918084-d6ec-433c-9fbe-8f1386fb347e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.299051 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.331287 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33918084-d6ec-433c-9fbe-8f1386fb347e" (UID: "33918084-d6ec-433c-9fbe-8f1386fb347e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.342543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data" (OuterVolumeSpecName: "config-data") pod "33918084-d6ec-433c-9fbe-8f1386fb347e" (UID: "33918084-d6ec-433c-9fbe-8f1386fb347e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382079 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382248 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382299 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx56n\" (UniqueName: \"kubernetes.io/projected/ba4d4a69-79d0-4f8b-83db-319fd0aca696-kube-api-access-xx56n\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382338 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382366 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382446 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382465 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382479 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npt7s\" (UniqueName: \"kubernetes.io/projected/33918084-d6ec-433c-9fbe-8f1386fb347e-kube-api-access-npt7s\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.382497 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33918084-d6ec-433c-9fbe-8f1386fb347e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.387114 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.387121 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.387270 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.387659 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba4d4a69-79d0-4f8b-83db-319fd0aca696-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.404549 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx56n\" (UniqueName: \"kubernetes.io/projected/ba4d4a69-79d0-4f8b-83db-319fd0aca696-kube-api-access-xx56n\") pod \"nova-cell1-novncproxy-0\" (UID: \"ba4d4a69-79d0-4f8b-83db-319fd0aca696\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.598713 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.929004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ktt7n" event={"ID":"33918084-d6ec-433c-9fbe-8f1386fb347e","Type":"ContainerDied","Data":"b337068ea6f92662254e301783f10c6f51ee25ffa0267f6084bf2edd43f07824"} Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.929466 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b337068ea6f92662254e301783f10c6f51ee25ffa0267f6084bf2edd43f07824" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.929531 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ktt7n" Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.945286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerStarted","Data":"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3"} Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.945366 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerStarted","Data":"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed"} Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.945387 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerStarted","Data":"f76fa665ad2d0e617e3039f7442d9f9d6b63011ebac299c8d8f38b5d3f73fdfe"} Oct 01 09:08:48 crc kubenswrapper[4837]: I1001 09:08:48.979315 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.979285756 podStartE2EDuration="2.979285756s" podCreationTimestamp="2025-10-01 09:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:48.964105709 +0000 UTC m=+7385.805713204" watchObservedRunningTime="2025-10-01 09:08:48.979285756 +0000 UTC m=+7385.820893221" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.058526 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.232670 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.233845 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-log" containerID="cri-o://4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" gracePeriod=30 Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.235074 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-api" containerID="cri-o://a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" gracePeriod=30 Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.248629 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.249073 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="02d99b3a-47b2-4e39-b144-7659c300cc25" containerName="nova-scheduler-scheduler" containerID="cri-o://eca879cf191100d8cde22e388f93e7afd846fe85c9cbddef53a6d3eb9fa32589" gracePeriod=30 Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.274488 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.328268 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.426671 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data\") pod \"cafbdbfe-2005-402e-a731-20f4afb49879\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.428296 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle\") pod \"cafbdbfe-2005-402e-a731-20f4afb49879\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.428456 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdcjb\" (UniqueName: \"kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb\") pod \"cafbdbfe-2005-402e-a731-20f4afb49879\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.428560 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts\") pod \"cafbdbfe-2005-402e-a731-20f4afb49879\" (UID: \"cafbdbfe-2005-402e-a731-20f4afb49879\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.448581 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb" (OuterVolumeSpecName: "kube-api-access-cdcjb") pod "cafbdbfe-2005-402e-a731-20f4afb49879" (UID: "cafbdbfe-2005-402e-a731-20f4afb49879"). InnerVolumeSpecName "kube-api-access-cdcjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.454819 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data" (OuterVolumeSpecName: "config-data") pod "cafbdbfe-2005-402e-a731-20f4afb49879" (UID: "cafbdbfe-2005-402e-a731-20f4afb49879"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.463438 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts" (OuterVolumeSpecName: "scripts") pod "cafbdbfe-2005-402e-a731-20f4afb49879" (UID: "cafbdbfe-2005-402e-a731-20f4afb49879"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.466253 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cafbdbfe-2005-402e-a731-20f4afb49879" (UID: "cafbdbfe-2005-402e-a731-20f4afb49879"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.532414 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdcjb\" (UniqueName: \"kubernetes.io/projected/cafbdbfe-2005-402e-a731-20f4afb49879-kube-api-access-cdcjb\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.532451 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.532462 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.532472 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cafbdbfe-2005-402e-a731-20f4afb49879-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.827265 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.828180 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6592600a-c30d-42ed-854a-22f489159c0e" path="/var/lib/kubelet/pods/6592600a-c30d-42ed-854a-22f489159c0e/volumes" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.937898 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle\") pod \"33353b74-a504-460b-8599-1d1e602b82a5\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.937962 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs\") pod \"33353b74-a504-460b-8599-1d1e602b82a5\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.938068 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlq5r\" (UniqueName: \"kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r\") pod \"33353b74-a504-460b-8599-1d1e602b82a5\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.938164 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data\") pod \"33353b74-a504-460b-8599-1d1e602b82a5\" (UID: \"33353b74-a504-460b-8599-1d1e602b82a5\") " Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.939464 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs" (OuterVolumeSpecName: "logs") pod "33353b74-a504-460b-8599-1d1e602b82a5" (UID: "33353b74-a504-460b-8599-1d1e602b82a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.958586 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r" (OuterVolumeSpecName: "kube-api-access-dlq5r") pod "33353b74-a504-460b-8599-1d1e602b82a5" (UID: "33353b74-a504-460b-8599-1d1e602b82a5"). InnerVolumeSpecName "kube-api-access-dlq5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.975173 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.975212 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pj4qg" event={"ID":"cafbdbfe-2005-402e-a731-20f4afb49879","Type":"ContainerDied","Data":"7b84e63f029563bc09930990a51d04d4ffef05e15e529ace168c9b9a0db1d703"} Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.975270 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b84e63f029563bc09930990a51d04d4ffef05e15e529ace168c9b9a0db1d703" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978603 4837 generic.go:334] "Generic (PLEG): container finished" podID="33353b74-a504-460b-8599-1d1e602b82a5" containerID="a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" exitCode=0 Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978671 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978711 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerDied","Data":"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b"} Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978736 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33353b74-a504-460b-8599-1d1e602b82a5" (UID: "33353b74-a504-460b-8599-1d1e602b82a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978766 4837 scope.go:117] "RemoveContainer" containerID="a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978643 4837 generic.go:334] "Generic (PLEG): container finished" podID="33353b74-a504-460b-8599-1d1e602b82a5" containerID="4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" exitCode=143 Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978748 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerDied","Data":"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681"} Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.978885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"33353b74-a504-460b-8599-1d1e602b82a5","Type":"ContainerDied","Data":"a88f70f8f05876ac908b5eff91e28cba052a2c86584cc4051ea28756795cba8c"} Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.983640 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ba4d4a69-79d0-4f8b-83db-319fd0aca696","Type":"ContainerStarted","Data":"e01cb9fceb08098cf994f7b2a3afae44fe720766649462a9b12522b3dddb5720"} Oct 01 09:08:49 crc kubenswrapper[4837]: I1001 09:08:49.983681 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ba4d4a69-79d0-4f8b-83db-319fd0aca696","Type":"ContainerStarted","Data":"1887f9357c7c86c0809ba842facb464ca309a26effa9f996dc88b8fb991c884a"} Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.017466 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data" (OuterVolumeSpecName: "config-data") pod "33353b74-a504-460b-8599-1d1e602b82a5" (UID: "33353b74-a504-460b-8599-1d1e602b82a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.030415 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.030386468 podStartE2EDuration="2.030386468s" podCreationTimestamp="2025-10-01 09:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:50.005167184 +0000 UTC m=+7386.846774679" watchObservedRunningTime="2025-10-01 09:08:50.030386468 +0000 UTC m=+7386.871993933" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.041757 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.041797 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33353b74-a504-460b-8599-1d1e602b82a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.041810 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33353b74-a504-460b-8599-1d1e602b82a5-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.041822 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlq5r\" (UniqueName: \"kubernetes.io/projected/33353b74-a504-460b-8599-1d1e602b82a5-kube-api-access-dlq5r\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.045251 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wnz4p"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.058964 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wnz4p"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.066777 4837 scope.go:117] "RemoveContainer" containerID="4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.088471 4837 scope.go:117] "RemoveContainer" containerID="a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" Oct 01 09:08:50 crc kubenswrapper[4837]: E1001 09:08:50.089135 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b\": container with ID starting with a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b not found: ID does not exist" containerID="a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.089174 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b"} err="failed to get container status \"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b\": rpc error: code = NotFound desc = could not find container \"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b\": container with ID starting with a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b not found: ID does not exist" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.089201 4837 scope.go:117] "RemoveContainer" containerID="4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" Oct 01 09:08:50 crc kubenswrapper[4837]: E1001 09:08:50.090357 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681\": container with ID starting with 4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681 not found: ID does not exist" containerID="4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.090396 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681"} err="failed to get container status \"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681\": rpc error: code = NotFound desc = could not find container \"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681\": container with ID starting with 4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681 not found: ID does not exist" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.090425 4837 scope.go:117] "RemoveContainer" containerID="a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.090857 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b"} err="failed to get container status \"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b\": rpc error: code = NotFound desc = could not find container \"a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b\": container with ID starting with a357a166e878cdce0c4ae5eeb9f28b76dff34864bb137334e171713588eda15b not found: ID does not exist" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.090907 4837 scope.go:117] "RemoveContainer" containerID="4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.091201 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681"} err="failed to get container status \"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681\": rpc error: code = NotFound desc = could not find container \"4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681\": container with ID starting with 4b26c74ad953701490e2aa50f6a866f11a7ea8ecb3b02ed7fd8fd9af51d1e681 not found: ID does not exist" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.322672 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.347468 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.356815 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: E1001 09:08:50.357404 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafbdbfe-2005-402e-a731-20f4afb49879" containerName="nova-cell1-conductor-db-sync" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357430 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafbdbfe-2005-402e-a731-20f4afb49879" containerName="nova-cell1-conductor-db-sync" Oct 01 09:08:50 crc kubenswrapper[4837]: E1001 09:08:50.357452 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-log" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357469 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-log" Oct 01 09:08:50 crc kubenswrapper[4837]: E1001 09:08:50.357492 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-api" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357506 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-api" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357853 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-log" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357881 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="33353b74-a504-460b-8599-1d1e602b82a5" containerName="nova-api-api" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.357911 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cafbdbfe-2005-402e-a731-20f4afb49879" containerName="nova-cell1-conductor-db-sync" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.359639 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.362340 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.365605 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.427114 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.429575 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.432113 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.441169 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.453210 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.453255 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k796\" (UniqueName: \"kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.453450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.453511 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kls5d\" (UniqueName: \"kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.453582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.454717 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.455210 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557259 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557301 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k796\" (UniqueName: \"kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557350 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557373 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kls5d\" (UniqueName: \"kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557395 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.557482 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.558438 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.562763 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.563803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.564227 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.565344 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.574800 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k796\" (UniqueName: \"kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796\") pod \"nova-cell1-conductor-0\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.594184 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kls5d\" (UniqueName: \"kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d\") pod \"nova-api-0\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.685735 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.745017 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.997318 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-log" containerID="cri-o://d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" gracePeriod=30 Oct 01 09:08:50 crc kubenswrapper[4837]: I1001 09:08:50.997506 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-metadata" containerID="cri-o://ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" gracePeriod=30 Oct 01 09:08:51 crc kubenswrapper[4837]: W1001 09:08:51.253370 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbcd4e9a_eedb_4ef6_8d59_e6db2849a7eb.slice/crio-ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80 WatchSource:0}: Error finding container ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80: Status 404 returned error can't find the container with id ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80 Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.258812 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.370560 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:08:51 crc kubenswrapper[4837]: W1001 09:08:51.389614 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d4b9891_1b5e_42d3_888f_2558f018e4b2.slice/crio-d695359fff4d55a1a3593c5d3d44cdf94e16f74908e7c27a9a59b3b56b5bd923 WatchSource:0}: Error finding container d695359fff4d55a1a3593c5d3d44cdf94e16f74908e7c27a9a59b3b56b5bd923: Status 404 returned error can't find the container with id d695359fff4d55a1a3593c5d3d44cdf94e16f74908e7c27a9a59b3b56b5bd923 Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.454992 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.535286 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.535551 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="dnsmasq-dns" containerID="cri-o://6e445beb75794c28ddc18870ed5e310001d0e1740b7ab1256fbf8d17b972c39f" gracePeriod=10 Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.744988 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.791543 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs\") pod \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.791603 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkfj7\" (UniqueName: \"kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7\") pod \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.791680 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle\") pod \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.791821 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data\") pod \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.791928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs\") pod \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\" (UID: \"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9\") " Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.792496 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs" (OuterVolumeSpecName: "logs") pod "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" (UID: "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.793676 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.799022 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7" (OuterVolumeSpecName: "kube-api-access-kkfj7") pod "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" (UID: "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9"). InnerVolumeSpecName "kube-api-access-kkfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.824013 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data" (OuterVolumeSpecName: "config-data") pod "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" (UID: "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.829182 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" (UID: "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.842490 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="186b2215-73a0-4142-b1d3-139f6b72a2b1" path="/var/lib/kubelet/pods/186b2215-73a0-4142-b1d3-139f6b72a2b1/volumes" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.843234 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33353b74-a504-460b-8599-1d1e602b82a5" path="/var/lib/kubelet/pods/33353b74-a504-460b-8599-1d1e602b82a5/volumes" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.855856 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" (UID: "00fb0081-d5ce-4392-bdd2-cb0642cfb7b9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.900596 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.900680 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkfj7\" (UniqueName: \"kubernetes.io/projected/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-kube-api-access-kkfj7\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.900744 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:51 crc kubenswrapper[4837]: I1001 09:08:51.900757 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.007636 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d4b9891-1b5e-42d3-888f-2558f018e4b2","Type":"ContainerStarted","Data":"0ea8b0ca63d78c0bcc8507109cd8380044e1ffa354eeb56d47eeea2b6a5133b6"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.007718 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d4b9891-1b5e-42d3-888f-2558f018e4b2","Type":"ContainerStarted","Data":"d695359fff4d55a1a3593c5d3d44cdf94e16f74908e7c27a9a59b3b56b5bd923"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.008031 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.010773 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerStarted","Data":"daafde44527af8306a7caeff258126a05924fc5ce3cdf085d25aee55f3bdf7b4"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.010835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerStarted","Data":"b89b918c7b46211af038bbb7d58bf08e5181f26962ca0b4b0a461a72f4a0a8ac"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.010855 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerStarted","Data":"ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013843 4837 generic.go:334] "Generic (PLEG): container finished" podID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerID="ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" exitCode=0 Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013877 4837 generic.go:334] "Generic (PLEG): container finished" podID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerID="d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" exitCode=143 Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013929 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerDied","Data":"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013956 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerDied","Data":"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013970 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"00fb0081-d5ce-4392-bdd2-cb0642cfb7b9","Type":"ContainerDied","Data":"f76fa665ad2d0e617e3039f7442d9f9d6b63011ebac299c8d8f38b5d3f73fdfe"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.013990 4837 scope.go:117] "RemoveContainer" containerID="ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.014120 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.019752 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerID="6e445beb75794c28ddc18870ed5e310001d0e1740b7ab1256fbf8d17b972c39f" exitCode=0 Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.019798 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" event={"ID":"c2dc9d74-5044-4242-bf5e-8b5221d0b143","Type":"ContainerDied","Data":"6e445beb75794c28ddc18870ed5e310001d0e1740b7ab1256fbf8d17b972c39f"} Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.033945 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.033925159 podStartE2EDuration="2.033925159s" podCreationTimestamp="2025-10-01 09:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:52.024134126 +0000 UTC m=+7388.865741601" watchObservedRunningTime="2025-10-01 09:08:52.033925159 +0000 UTC m=+7388.875532624" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.052599 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.052577731 podStartE2EDuration="2.052577731s" podCreationTimestamp="2025-10-01 09:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:52.047884295 +0000 UTC m=+7388.889491750" watchObservedRunningTime="2025-10-01 09:08:52.052577731 +0000 UTC m=+7388.894185206" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.061726 4837 scope.go:117] "RemoveContainer" containerID="d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.085399 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.118815 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.124815 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:52 crc kubenswrapper[4837]: E1001 09:08:52.125293 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-metadata" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.125310 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-metadata" Oct 01 09:08:52 crc kubenswrapper[4837]: E1001 09:08:52.125358 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-log" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.125367 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-log" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.125553 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-metadata" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.125568 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" containerName="nova-metadata-log" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.126559 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.126845 4837 scope.go:117] "RemoveContainer" containerID="ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" Oct 01 09:08:52 crc kubenswrapper[4837]: E1001 09:08:52.128264 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3\": container with ID starting with ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3 not found: ID does not exist" containerID="ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.128310 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3"} err="failed to get container status \"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3\": rpc error: code = NotFound desc = could not find container \"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3\": container with ID starting with ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3 not found: ID does not exist" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.128335 4837 scope.go:117] "RemoveContainer" containerID="d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" Oct 01 09:08:52 crc kubenswrapper[4837]: E1001 09:08:52.128845 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed\": container with ID starting with d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed not found: ID does not exist" containerID="d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.128901 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed"} err="failed to get container status \"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed\": rpc error: code = NotFound desc = could not find container \"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed\": container with ID starting with d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed not found: ID does not exist" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.128942 4837 scope.go:117] "RemoveContainer" containerID="ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.132475 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3"} err="failed to get container status \"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3\": rpc error: code = NotFound desc = could not find container \"ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3\": container with ID starting with ca111d5c8b4efb3e73aca9f70505404a9ebbf024b929c2db78022674a0e086e3 not found: ID does not exist" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.132575 4837 scope.go:117] "RemoveContainer" containerID="d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.133023 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.133318 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.134794 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed"} err="failed to get container status \"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed\": rpc error: code = NotFound desc = could not find container \"d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed\": container with ID starting with d0bafedd998d7830de88f73432b0dd4ff5101e5ecf16603bf17996efe496cbed not found: ID does not exist" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.142975 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.206061 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pb4j\" (UniqueName: \"kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.206230 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.206277 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.206368 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.206529 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.273798 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.312496 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb\") pod \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.312638 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config\") pod \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.312758 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc\") pod \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.312831 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb\") pod \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.312865 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gbjg\" (UniqueName: \"kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg\") pod \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\" (UID: \"c2dc9d74-5044-4242-bf5e-8b5221d0b143\") " Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314010 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314086 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pb4j\" (UniqueName: \"kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314202 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314614 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314675 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.314502 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.322611 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg" (OuterVolumeSpecName: "kube-api-access-4gbjg") pod "c2dc9d74-5044-4242-bf5e-8b5221d0b143" (UID: "c2dc9d74-5044-4242-bf5e-8b5221d0b143"). InnerVolumeSpecName "kube-api-access-4gbjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.327818 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.327979 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.333372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pb4j\" (UniqueName: \"kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.363593 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.388146 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config" (OuterVolumeSpecName: "config") pod "c2dc9d74-5044-4242-bf5e-8b5221d0b143" (UID: "c2dc9d74-5044-4242-bf5e-8b5221d0b143"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.394937 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2dc9d74-5044-4242-bf5e-8b5221d0b143" (UID: "c2dc9d74-5044-4242-bf5e-8b5221d0b143"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.404108 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2dc9d74-5044-4242-bf5e-8b5221d0b143" (UID: "c2dc9d74-5044-4242-bf5e-8b5221d0b143"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.405662 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2dc9d74-5044-4242-bf5e-8b5221d0b143" (UID: "c2dc9d74-5044-4242-bf5e-8b5221d0b143"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.419647 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.419671 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.419682 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.419704 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gbjg\" (UniqueName: \"kubernetes.io/projected/c2dc9d74-5044-4242-bf5e-8b5221d0b143-kube-api-access-4gbjg\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.419714 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dc9d74-5044-4242-bf5e-8b5221d0b143-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.447011 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:08:52 crc kubenswrapper[4837]: I1001 09:08:52.994846 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.034213 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" event={"ID":"c2dc9d74-5044-4242-bf5e-8b5221d0b143","Type":"ContainerDied","Data":"8c3675672bcd78347d64674a7af45b7e588122b42dc62d137724dca1be911526"} Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.034259 4837 scope.go:117] "RemoveContainer" containerID="6e445beb75794c28ddc18870ed5e310001d0e1740b7ab1256fbf8d17b972c39f" Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.034296 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d58b9845-j5hrd" Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.040458 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerStarted","Data":"5a6253aaac73cd24267db54b7dd9fa9d9afc13b29126d8eceb4dcce88ea348b1"} Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.091021 4837 scope.go:117] "RemoveContainer" containerID="9926a8ade6c8d86b1ceb312826fc2b62482a8c6032e0bce19f1b7e9e0c8cb390" Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.197329 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.207403 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d58b9845-j5hrd"] Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.599882 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.835003 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00fb0081-d5ce-4392-bdd2-cb0642cfb7b9" path="/var/lib/kubelet/pods/00fb0081-d5ce-4392-bdd2-cb0642cfb7b9/volumes" Oct 01 09:08:53 crc kubenswrapper[4837]: I1001 09:08:53.835602 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" path="/var/lib/kubelet/pods/c2dc9d74-5044-4242-bf5e-8b5221d0b143/volumes" Oct 01 09:08:54 crc kubenswrapper[4837]: I1001 09:08:54.054325 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerStarted","Data":"00990d322d65cab6f5c21f980d551e418ccf62d5747dca9d385e50863ce1f874"} Oct 01 09:08:54 crc kubenswrapper[4837]: I1001 09:08:54.054386 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerStarted","Data":"d18644e064fa922c7cfb2d3a470fa39c19d2d3738633b7f5e9788fe24bfcbde3"} Oct 01 09:08:54 crc kubenswrapper[4837]: I1001 09:08:54.091296 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.091265572 podStartE2EDuration="2.091265572s" podCreationTimestamp="2025-10-01 09:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:08:54.077357027 +0000 UTC m=+7390.918964522" watchObservedRunningTime="2025-10-01 09:08:54.091265572 +0000 UTC m=+7390.932873057" Oct 01 09:08:57 crc kubenswrapper[4837]: I1001 09:08:57.447108 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:08:57 crc kubenswrapper[4837]: I1001 09:08:57.447601 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:08:58 crc kubenswrapper[4837]: I1001 09:08:58.600144 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:58 crc kubenswrapper[4837]: I1001 09:08:58.634169 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:08:59 crc kubenswrapper[4837]: I1001 09:08:59.142411 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:09:00 crc kubenswrapper[4837]: I1001 09:09:00.686776 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:09:00 crc kubenswrapper[4837]: I1001 09:09:00.687137 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:09:00 crc kubenswrapper[4837]: I1001 09:09:00.802440 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.408662 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-t9n5g"] Oct 01 09:09:01 crc kubenswrapper[4837]: E1001 09:09:01.409046 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="dnsmasq-dns" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.409061 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="dnsmasq-dns" Oct 01 09:09:01 crc kubenswrapper[4837]: E1001 09:09:01.409080 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="init" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.409086 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="init" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.409264 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dc9d74-5044-4242-bf5e-8b5221d0b143" containerName="dnsmasq-dns" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.409951 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.412809 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.414258 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.415436 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.415596 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.415657 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkvck\" (UniqueName: \"kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.415828 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.426514 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9n5g"] Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.516738 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.517165 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkvck\" (UniqueName: \"kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.517297 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.517500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.523941 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.524809 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.533794 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.537361 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkvck\" (UniqueName: \"kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck\") pod \"nova-cell1-cell-mapping-t9n5g\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.732544 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.769034 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.101:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:01 crc kubenswrapper[4837]: I1001 09:09:01.769071 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.101:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:02 crc kubenswrapper[4837]: I1001 09:09:02.231957 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9n5g"] Oct 01 09:09:02 crc kubenswrapper[4837]: W1001 09:09:02.234753 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb63f46a7_38b3_4529_bbc2_ea5665df205f.slice/crio-9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446 WatchSource:0}: Error finding container 9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446: Status 404 returned error can't find the container with id 9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446 Oct 01 09:09:02 crc kubenswrapper[4837]: I1001 09:09:02.447291 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:09:02 crc kubenswrapper[4837]: I1001 09:09:02.447902 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:09:03 crc kubenswrapper[4837]: I1001 09:09:03.168884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9n5g" event={"ID":"b63f46a7-38b3-4529-bbc2-ea5665df205f","Type":"ContainerStarted","Data":"cf80aa3de92278f5af049dcc2a1d5b6ff02a73f47367c24496960c65cf828ee7"} Oct 01 09:09:03 crc kubenswrapper[4837]: I1001 09:09:03.168974 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9n5g" event={"ID":"b63f46a7-38b3-4529-bbc2-ea5665df205f","Type":"ContainerStarted","Data":"9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446"} Oct 01 09:09:03 crc kubenswrapper[4837]: I1001 09:09:03.189838 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-t9n5g" podStartSLOduration=2.189816491 podStartE2EDuration="2.189816491s" podCreationTimestamp="2025-10-01 09:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:03.186539971 +0000 UTC m=+7400.028147426" watchObservedRunningTime="2025-10-01 09:09:03.189816491 +0000 UTC m=+7400.031423956" Oct 01 09:09:03 crc kubenswrapper[4837]: I1001 09:09:03.459006 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:03 crc kubenswrapper[4837]: I1001 09:09:03.459050 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.103:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:07 crc kubenswrapper[4837]: I1001 09:09:07.226404 4837 generic.go:334] "Generic (PLEG): container finished" podID="b63f46a7-38b3-4529-bbc2-ea5665df205f" containerID="cf80aa3de92278f5af049dcc2a1d5b6ff02a73f47367c24496960c65cf828ee7" exitCode=0 Oct 01 09:09:07 crc kubenswrapper[4837]: I1001 09:09:07.226540 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9n5g" event={"ID":"b63f46a7-38b3-4529-bbc2-ea5665df205f","Type":"ContainerDied","Data":"cf80aa3de92278f5af049dcc2a1d5b6ff02a73f47367c24496960c65cf828ee7"} Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.705984 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.894257 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkvck\" (UniqueName: \"kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck\") pod \"b63f46a7-38b3-4529-bbc2-ea5665df205f\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.894332 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts\") pod \"b63f46a7-38b3-4529-bbc2-ea5665df205f\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.894451 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle\") pod \"b63f46a7-38b3-4529-bbc2-ea5665df205f\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.894515 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data\") pod \"b63f46a7-38b3-4529-bbc2-ea5665df205f\" (UID: \"b63f46a7-38b3-4529-bbc2-ea5665df205f\") " Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.902945 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck" (OuterVolumeSpecName: "kube-api-access-gkvck") pod "b63f46a7-38b3-4529-bbc2-ea5665df205f" (UID: "b63f46a7-38b3-4529-bbc2-ea5665df205f"). InnerVolumeSpecName "kube-api-access-gkvck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.907260 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts" (OuterVolumeSpecName: "scripts") pod "b63f46a7-38b3-4529-bbc2-ea5665df205f" (UID: "b63f46a7-38b3-4529-bbc2-ea5665df205f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.947727 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b63f46a7-38b3-4529-bbc2-ea5665df205f" (UID: "b63f46a7-38b3-4529-bbc2-ea5665df205f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.948326 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data" (OuterVolumeSpecName: "config-data") pod "b63f46a7-38b3-4529-bbc2-ea5665df205f" (UID: "b63f46a7-38b3-4529-bbc2-ea5665df205f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.997798 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.997990 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkvck\" (UniqueName: \"kubernetes.io/projected/b63f46a7-38b3-4529-bbc2-ea5665df205f-kube-api-access-gkvck\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.998130 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:08 crc kubenswrapper[4837]: I1001 09:09:08.998259 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63f46a7-38b3-4529-bbc2-ea5665df205f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.258900 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t9n5g" event={"ID":"b63f46a7-38b3-4529-bbc2-ea5665df205f","Type":"ContainerDied","Data":"9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446"} Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.258954 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9380b8684668397576fbbf6a952fc0677b578b31e52fc6c80acc5d13d0a0d446" Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.259114 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t9n5g" Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.499428 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.500595 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-api" containerID="cri-o://daafde44527af8306a7caeff258126a05924fc5ce3cdf085d25aee55f3bdf7b4" gracePeriod=30 Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.501127 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-log" containerID="cri-o://b89b918c7b46211af038bbb7d58bf08e5181f26962ca0b4b0a461a72f4a0a8ac" gracePeriod=30 Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.543491 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.546105 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-log" containerID="cri-o://d18644e064fa922c7cfb2d3a470fa39c19d2d3738633b7f5e9788fe24bfcbde3" gracePeriod=30 Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.546499 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-metadata" containerID="cri-o://00990d322d65cab6f5c21f980d551e418ccf62d5747dca9d385e50863ce1f874" gracePeriod=30 Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.905004 4837 scope.go:117] "RemoveContainer" containerID="25d709f52e8b9e13ce453c52a7135b37f25934800aa0583b8ec5b2cc0a17f4aa" Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.947492 4837 scope.go:117] "RemoveContainer" containerID="1a7a4110f9731589a3b66021a05a5c4fa0b3cbc7e42312437b181b3af65f691c" Oct 01 09:09:09 crc kubenswrapper[4837]: I1001 09:09:09.988099 4837 scope.go:117] "RemoveContainer" containerID="ecb955d2f94ccc6c37f63bf596afd0d2e044080e0c1141afd28f384b7f7e2ed3" Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.051851 4837 scope.go:117] "RemoveContainer" containerID="26289cb6f00e22a8b002ea5b7e6fb3e1b300f8d23fa798f2de0896520262a3a7" Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.086299 4837 scope.go:117] "RemoveContainer" containerID="c3f4acfe0747d97132b8302af5952f78b17b79af690b8c297660fb212f1dad2a" Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.124845 4837 scope.go:117] "RemoveContainer" containerID="473552225d6a3ada480afee842621cdd2f341e1545cefea837cb137e985b3558" Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.276038 4837 generic.go:334] "Generic (PLEG): container finished" podID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerID="d18644e064fa922c7cfb2d3a470fa39c19d2d3738633b7f5e9788fe24bfcbde3" exitCode=143 Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.276169 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerDied","Data":"d18644e064fa922c7cfb2d3a470fa39c19d2d3738633b7f5e9788fe24bfcbde3"} Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.286035 4837 generic.go:334] "Generic (PLEG): container finished" podID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerID="b89b918c7b46211af038bbb7d58bf08e5181f26962ca0b4b0a461a72f4a0a8ac" exitCode=143 Oct 01 09:09:10 crc kubenswrapper[4837]: I1001 09:09:10.286144 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerDied","Data":"b89b918c7b46211af038bbb7d58bf08e5181f26962ca0b4b0a461a72f4a0a8ac"} Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.392774 4837 generic.go:334] "Generic (PLEG): container finished" podID="02d99b3a-47b2-4e39-b144-7659c300cc25" containerID="eca879cf191100d8cde22e388f93e7afd846fe85c9cbddef53a6d3eb9fa32589" exitCode=137 Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.392950 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02d99b3a-47b2-4e39-b144-7659c300cc25","Type":"ContainerDied","Data":"eca879cf191100d8cde22e388f93e7afd846fe85c9cbddef53a6d3eb9fa32589"} Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.818146 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.948410 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jctz\" (UniqueName: \"kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz\") pod \"02d99b3a-47b2-4e39-b144-7659c300cc25\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.948545 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data\") pod \"02d99b3a-47b2-4e39-b144-7659c300cc25\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.948569 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle\") pod \"02d99b3a-47b2-4e39-b144-7659c300cc25\" (UID: \"02d99b3a-47b2-4e39-b144-7659c300cc25\") " Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.954524 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz" (OuterVolumeSpecName: "kube-api-access-2jctz") pod "02d99b3a-47b2-4e39-b144-7659c300cc25" (UID: "02d99b3a-47b2-4e39-b144-7659c300cc25"). InnerVolumeSpecName "kube-api-access-2jctz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.974341 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02d99b3a-47b2-4e39-b144-7659c300cc25" (UID: "02d99b3a-47b2-4e39-b144-7659c300cc25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:19 crc kubenswrapper[4837]: I1001 09:09:19.980255 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data" (OuterVolumeSpecName: "config-data") pod "02d99b3a-47b2-4e39-b144-7659c300cc25" (UID: "02d99b3a-47b2-4e39-b144-7659c300cc25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.051617 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jctz\" (UniqueName: \"kubernetes.io/projected/02d99b3a-47b2-4e39-b144-7659c300cc25-kube-api-access-2jctz\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.051666 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.051685 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d99b3a-47b2-4e39-b144-7659c300cc25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.408394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02d99b3a-47b2-4e39-b144-7659c300cc25","Type":"ContainerDied","Data":"a57414b9b080d78233a94fc3b1d1a84b47507ec67ab9408b44b603206dd26569"} Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.408468 4837 scope.go:117] "RemoveContainer" containerID="eca879cf191100d8cde22e388f93e7afd846fe85c9cbddef53a6d3eb9fa32589" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.408500 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.457457 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.467408 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.500660 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:09:20 crc kubenswrapper[4837]: E1001 09:09:20.501280 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d99b3a-47b2-4e39-b144-7659c300cc25" containerName="nova-scheduler-scheduler" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.501313 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d99b3a-47b2-4e39-b144-7659c300cc25" containerName="nova-scheduler-scheduler" Oct 01 09:09:20 crc kubenswrapper[4837]: E1001 09:09:20.501340 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63f46a7-38b3-4529-bbc2-ea5665df205f" containerName="nova-manage" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.501353 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63f46a7-38b3-4529-bbc2-ea5665df205f" containerName="nova-manage" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.501722 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d99b3a-47b2-4e39-b144-7659c300cc25" containerName="nova-scheduler-scheduler" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.501762 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63f46a7-38b3-4529-bbc2-ea5665df205f" containerName="nova-manage" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.502880 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.511494 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.531682 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.662873 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.663007 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.663123 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7hnz\" (UniqueName: \"kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.685970 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.686283 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.765416 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.765580 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7hnz\" (UniqueName: \"kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.765728 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.775860 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.776273 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.789686 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7hnz\" (UniqueName: \"kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz\") pod \"nova-scheduler-0\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " pod="openstack/nova-scheduler-0" Oct 01 09:09:20 crc kubenswrapper[4837]: I1001 09:09:20.855926 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:09:21 crc kubenswrapper[4837]: I1001 09:09:21.397217 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:09:21 crc kubenswrapper[4837]: I1001 09:09:21.427436 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6","Type":"ContainerStarted","Data":"5d8fbbf64463382b95030bfdb0374eb92b780dc108b780a5082847332cff5003"} Oct 01 09:09:21 crc kubenswrapper[4837]: I1001 09:09:21.834241 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d99b3a-47b2-4e39-b144-7659c300cc25" path="/var/lib/kubelet/pods/02d99b3a-47b2-4e39-b144-7659c300cc25/volumes" Oct 01 09:09:22 crc kubenswrapper[4837]: I1001 09:09:22.442918 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6","Type":"ContainerStarted","Data":"e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e"} Oct 01 09:09:22 crc kubenswrapper[4837]: I1001 09:09:22.472386 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.472357792 podStartE2EDuration="2.472357792s" podCreationTimestamp="2025-10-01 09:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:22.458943331 +0000 UTC m=+7419.300550826" watchObservedRunningTime="2025-10-01 09:09:22.472357792 +0000 UTC m=+7419.313965287" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.458163 4837 generic.go:334] "Generic (PLEG): container finished" podID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerID="daafde44527af8306a7caeff258126a05924fc5ce3cdf085d25aee55f3bdf7b4" exitCode=0 Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.458282 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerDied","Data":"daafde44527af8306a7caeff258126a05924fc5ce3cdf085d25aee55f3bdf7b4"} Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.458656 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb","Type":"ContainerDied","Data":"ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80"} Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.458677 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba28df68f535687c91cda8d546c88f3b6a0ebd01876509c83047cc74117b7e80" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.466914 4837 generic.go:334] "Generic (PLEG): container finished" podID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerID="00990d322d65cab6f5c21f980d551e418ccf62d5747dca9d385e50863ce1f874" exitCode=0 Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.467017 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerDied","Data":"00990d322d65cab6f5c21f980d551e418ccf62d5747dca9d385e50863ce1f874"} Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.467099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0ed03a31-8838-4d89-9449-c31bbef06bd4","Type":"ContainerDied","Data":"5a6253aaac73cd24267db54b7dd9fa9d9afc13b29126d8eceb4dcce88ea348b1"} Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.467130 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a6253aaac73cd24267db54b7dd9fa9d9afc13b29126d8eceb4dcce88ea348b1" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.480582 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.491753 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.636906 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data\") pod \"0ed03a31-8838-4d89-9449-c31bbef06bd4\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.636994 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs\") pod \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637021 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pb4j\" (UniqueName: \"kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j\") pod \"0ed03a31-8838-4d89-9449-c31bbef06bd4\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637039 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs\") pod \"0ed03a31-8838-4d89-9449-c31bbef06bd4\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637146 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data\") pod \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637167 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle\") pod \"0ed03a31-8838-4d89-9449-c31bbef06bd4\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637224 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle\") pod \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637242 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs\") pod \"0ed03a31-8838-4d89-9449-c31bbef06bd4\" (UID: \"0ed03a31-8838-4d89-9449-c31bbef06bd4\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637273 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kls5d\" (UniqueName: \"kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d\") pod \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\" (UID: \"fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb\") " Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.637835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs" (OuterVolumeSpecName: "logs") pod "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" (UID: "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.638578 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs" (OuterVolumeSpecName: "logs") pod "0ed03a31-8838-4d89-9449-c31bbef06bd4" (UID: "0ed03a31-8838-4d89-9449-c31bbef06bd4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.641952 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d" (OuterVolumeSpecName: "kube-api-access-kls5d") pod "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" (UID: "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb"). InnerVolumeSpecName "kube-api-access-kls5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.648932 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j" (OuterVolumeSpecName: "kube-api-access-2pb4j") pod "0ed03a31-8838-4d89-9449-c31bbef06bd4" (UID: "0ed03a31-8838-4d89-9449-c31bbef06bd4"). InnerVolumeSpecName "kube-api-access-2pb4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.665332 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ed03a31-8838-4d89-9449-c31bbef06bd4" (UID: "0ed03a31-8838-4d89-9449-c31bbef06bd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.670037 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" (UID: "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.671998 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data" (OuterVolumeSpecName: "config-data") pod "0ed03a31-8838-4d89-9449-c31bbef06bd4" (UID: "0ed03a31-8838-4d89-9449-c31bbef06bd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.677375 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data" (OuterVolumeSpecName: "config-data") pod "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" (UID: "fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.692242 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0ed03a31-8838-4d89-9449-c31bbef06bd4" (UID: "0ed03a31-8838-4d89-9449-c31bbef06bd4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739442 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739524 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739553 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pb4j\" (UniqueName: \"kubernetes.io/projected/0ed03a31-8838-4d89-9449-c31bbef06bd4-kube-api-access-2pb4j\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739570 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739578 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739586 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed03a31-8838-4d89-9449-c31bbef06bd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739595 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739602 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ed03a31-8838-4d89-9449-c31bbef06bd4-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:23 crc kubenswrapper[4837]: I1001 09:09:23.739610 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kls5d\" (UniqueName: \"kubernetes.io/projected/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb-kube-api-access-kls5d\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.478626 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.478770 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.516797 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.531588 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.549872 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.565821 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.593614 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: E1001 09:09:24.594181 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-log" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594204 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-log" Oct 01 09:09:24 crc kubenswrapper[4837]: E1001 09:09:24.594222 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-log" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594229 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-log" Oct 01 09:09:24 crc kubenswrapper[4837]: E1001 09:09:24.594247 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-metadata" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594254 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-metadata" Oct 01 09:09:24 crc kubenswrapper[4837]: E1001 09:09:24.594285 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-api" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594292 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-api" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594494 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-metadata" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594520 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-api" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594539 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" containerName="nova-metadata-log" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.594546 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" containerName="nova-api-log" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.596342 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.606734 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.609097 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.612771 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.612990 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.613145 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.615802 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.623763 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.759639 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.760121 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.760310 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbbzw\" (UniqueName: \"kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.760592 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.760779 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5p5\" (UniqueName: \"kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.760985 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.761156 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.761327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.761567 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.862970 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863029 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863059 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863119 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863165 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863190 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863211 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbbzw\" (UniqueName: \"kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863259 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5p5\" (UniqueName: \"kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.863543 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.864580 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.873932 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.874818 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.875525 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.875805 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.878838 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.884906 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbbzw\" (UniqueName: \"kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw\") pod \"nova-metadata-0\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.887495 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5p5\" (UniqueName: \"kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5\") pod \"nova-api-0\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " pod="openstack/nova-api-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.940612 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:09:24 crc kubenswrapper[4837]: I1001 09:09:24.949499 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.460092 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.493030 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerStarted","Data":"8c9e3b226e3af16c08072a3b0e6d47271ebf5f0829acf0db676829ae712d9d67"} Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.540467 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:09:25 crc kubenswrapper[4837]: W1001 09:09:25.554997 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47457bb8_964e_452f_886f_9019dd2f818a.slice/crio-50bf945ea109e52943eac336fa48d034a14ff440fe8280e9eed8a4605c365c8a WatchSource:0}: Error finding container 50bf945ea109e52943eac336fa48d034a14ff440fe8280e9eed8a4605c365c8a: Status 404 returned error can't find the container with id 50bf945ea109e52943eac336fa48d034a14ff440fe8280e9eed8a4605c365c8a Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.831905 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed03a31-8838-4d89-9449-c31bbef06bd4" path="/var/lib/kubelet/pods/0ed03a31-8838-4d89-9449-c31bbef06bd4/volumes" Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.833471 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb" path="/var/lib/kubelet/pods/fbcd4e9a-eedb-4ef6-8d59-e6db2849a7eb/volumes" Oct 01 09:09:25 crc kubenswrapper[4837]: I1001 09:09:25.856519 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.510118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerStarted","Data":"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826"} Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.510483 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerStarted","Data":"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f"} Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.510495 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerStarted","Data":"50bf945ea109e52943eac336fa48d034a14ff440fe8280e9eed8a4605c365c8a"} Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.513102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerStarted","Data":"9d4fd1a9c32665c0d661aac66a9c2d2778490414ea635f9e03b418d0cf33c8f9"} Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.513154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerStarted","Data":"baf2e4bc5600fceeee3a4975a44a7d65f3ce1145336e48b285792263e91ccdd8"} Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.529512 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.529490074 podStartE2EDuration="2.529490074s" podCreationTimestamp="2025-10-01 09:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:26.527398573 +0000 UTC m=+7423.369006038" watchObservedRunningTime="2025-10-01 09:09:26.529490074 +0000 UTC m=+7423.371097529" Oct 01 09:09:26 crc kubenswrapper[4837]: I1001 09:09:26.560925 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.560906412 podStartE2EDuration="2.560906412s" podCreationTimestamp="2025-10-01 09:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:26.55314824 +0000 UTC m=+7423.394755705" watchObservedRunningTime="2025-10-01 09:09:26.560906412 +0000 UTC m=+7423.402513867" Oct 01 09:09:29 crc kubenswrapper[4837]: I1001 09:09:29.942597 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:09:29 crc kubenswrapper[4837]: I1001 09:09:29.943081 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:09:30 crc kubenswrapper[4837]: I1001 09:09:30.856599 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 09:09:30 crc kubenswrapper[4837]: I1001 09:09:30.885838 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 09:09:31 crc kubenswrapper[4837]: I1001 09:09:31.622407 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 09:09:34 crc kubenswrapper[4837]: I1001 09:09:34.941902 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:09:34 crc kubenswrapper[4837]: I1001 09:09:34.942390 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:09:34 crc kubenswrapper[4837]: I1001 09:09:34.949705 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:09:34 crc kubenswrapper[4837]: I1001 09:09:34.950144 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:09:35 crc kubenswrapper[4837]: I1001 09:09:35.955868 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.106:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:35 crc kubenswrapper[4837]: I1001 09:09:35.955883 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.106:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:36 crc kubenswrapper[4837]: I1001 09:09:36.038023 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.107:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:36 crc kubenswrapper[4837]: I1001 09:09:36.038059 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.107:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.960462 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.961725 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.962281 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.962305 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.964153 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.965497 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.966933 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.969266 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:09:44 crc kubenswrapper[4837]: I1001 09:09:44.969995 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.172457 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.174394 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.204858 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.228314 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.228393 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdfhp\" (UniqueName: \"kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.228429 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.228487 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.228522 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.329864 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdfhp\" (UniqueName: \"kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.329916 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.329963 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.329993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.330066 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.330962 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.330983 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.330980 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.331529 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.361863 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdfhp\" (UniqueName: \"kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp\") pod \"dnsmasq-dns-57b74c447c-79rbh\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:45 crc kubenswrapper[4837]: I1001 09:09:45.509543 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:46 crc kubenswrapper[4837]: I1001 09:09:45.754680 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:09:46 crc kubenswrapper[4837]: I1001 09:09:46.587149 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:09:46 crc kubenswrapper[4837]: W1001 09:09:46.606923 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aa0abb8_7138_4cc0_9c27_3cbf25a6bd09.slice/crio-467bbdbb1cc8353a2cc109d148ec841ec8df4fe7288c9c890dcbddd67acbae79 WatchSource:0}: Error finding container 467bbdbb1cc8353a2cc109d148ec841ec8df4fe7288c9c890dcbddd67acbae79: Status 404 returned error can't find the container with id 467bbdbb1cc8353a2cc109d148ec841ec8df4fe7288c9c890dcbddd67acbae79 Oct 01 09:09:46 crc kubenswrapper[4837]: I1001 09:09:46.758567 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" event={"ID":"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09","Type":"ContainerStarted","Data":"467bbdbb1cc8353a2cc109d148ec841ec8df4fe7288c9c890dcbddd67acbae79"} Oct 01 09:09:47 crc kubenswrapper[4837]: I1001 09:09:47.799613 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" event={"ID":"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09","Type":"ContainerDied","Data":"c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6"} Oct 01 09:09:47 crc kubenswrapper[4837]: I1001 09:09:47.799527 4837 generic.go:334] "Generic (PLEG): container finished" podID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerID="c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6" exitCode=0 Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.199513 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.200004 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-log" containerID="cri-o://baf2e4bc5600fceeee3a4975a44a7d65f3ce1145336e48b285792263e91ccdd8" gracePeriod=30 Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.200140 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-api" containerID="cri-o://9d4fd1a9c32665c0d661aac66a9c2d2778490414ea635f9e03b418d0cf33c8f9" gracePeriod=30 Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.810334 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" event={"ID":"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09","Type":"ContainerStarted","Data":"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f"} Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.811026 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.812176 4837 generic.go:334] "Generic (PLEG): container finished" podID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerID="baf2e4bc5600fceeee3a4975a44a7d65f3ce1145336e48b285792263e91ccdd8" exitCode=143 Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.812234 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerDied","Data":"baf2e4bc5600fceeee3a4975a44a7d65f3ce1145336e48b285792263e91ccdd8"} Oct 01 09:09:48 crc kubenswrapper[4837]: I1001 09:09:48.839587 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" podStartSLOduration=3.839566648 podStartE2EDuration="3.839566648s" podCreationTimestamp="2025-10-01 09:09:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:48.829930629 +0000 UTC m=+7445.671538094" watchObservedRunningTime="2025-10-01 09:09:48.839566648 +0000 UTC m=+7445.681174113" Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.846293 4837 generic.go:334] "Generic (PLEG): container finished" podID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerID="9d4fd1a9c32665c0d661aac66a9c2d2778490414ea635f9e03b418d0cf33c8f9" exitCode=0 Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.846784 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerDied","Data":"9d4fd1a9c32665c0d661aac66a9c2d2778490414ea635f9e03b418d0cf33c8f9"} Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.846827 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ddf126bf-f08d-4da5-9c78-28dc2e27d157","Type":"ContainerDied","Data":"8c9e3b226e3af16c08072a3b0e6d47271ebf5f0829acf0db676829ae712d9d67"} Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.846848 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c9e3b226e3af16c08072a3b0e6d47271ebf5f0829acf0db676829ae712d9d67" Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.859499 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.987590 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk5p5\" (UniqueName: \"kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5\") pod \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.987932 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data\") pod \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.988106 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle\") pod \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.988167 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs\") pod \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\" (UID: \"ddf126bf-f08d-4da5-9c78-28dc2e27d157\") " Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.989063 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs" (OuterVolumeSpecName: "logs") pod "ddf126bf-f08d-4da5-9c78-28dc2e27d157" (UID: "ddf126bf-f08d-4da5-9c78-28dc2e27d157"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:09:51 crc kubenswrapper[4837]: I1001 09:09:51.998247 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5" (OuterVolumeSpecName: "kube-api-access-dk5p5") pod "ddf126bf-f08d-4da5-9c78-28dc2e27d157" (UID: "ddf126bf-f08d-4da5-9c78-28dc2e27d157"). InnerVolumeSpecName "kube-api-access-dk5p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.020990 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data" (OuterVolumeSpecName: "config-data") pod "ddf126bf-f08d-4da5-9c78-28dc2e27d157" (UID: "ddf126bf-f08d-4da5-9c78-28dc2e27d157"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.025741 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf126bf-f08d-4da5-9c78-28dc2e27d157" (UID: "ddf126bf-f08d-4da5-9c78-28dc2e27d157"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.092052 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.092122 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddf126bf-f08d-4da5-9c78-28dc2e27d157-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.092140 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk5p5\" (UniqueName: \"kubernetes.io/projected/ddf126bf-f08d-4da5-9c78-28dc2e27d157-kube-api-access-dk5p5\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.092154 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf126bf-f08d-4da5-9c78-28dc2e27d157-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.853853 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.884486 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.892826 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.914251 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:52 crc kubenswrapper[4837]: E1001 09:09:52.914614 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-api" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.914626 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-api" Oct 01 09:09:52 crc kubenswrapper[4837]: E1001 09:09:52.914653 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-log" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.914662 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-log" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.914847 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-log" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.914867 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" containerName="nova-api-api" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.915783 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.919166 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.919594 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.919759 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 09:09:52 crc kubenswrapper[4837]: I1001 09:09:52.923571 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012073 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hnm7\" (UniqueName: \"kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012366 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012394 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012427 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012682 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.012879 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.114873 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hnm7\" (UniqueName: \"kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.115100 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.115180 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.115272 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.115424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.115502 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.116079 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.121162 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.121657 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.122492 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.123016 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.140143 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hnm7\" (UniqueName: \"kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7\") pod \"nova-api-0\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.246723 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.555361 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.855364 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf126bf-f08d-4da5-9c78-28dc2e27d157" path="/var/lib/kubelet/pods/ddf126bf-f08d-4da5-9c78-28dc2e27d157/volumes" Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.870562 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerStarted","Data":"79e2f40b5f6c2e364658462e4dfd2f3a4aadc83ed2cf052842724dcded0d8b05"} Oct 01 09:09:53 crc kubenswrapper[4837]: I1001 09:09:53.870613 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerStarted","Data":"8feeaca13a24c80d920f836351ca0430bb33ea168710af63916654deff1ad1e8"} Oct 01 09:09:54 crc kubenswrapper[4837]: I1001 09:09:54.883620 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerStarted","Data":"f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2"} Oct 01 09:09:54 crc kubenswrapper[4837]: I1001 09:09:54.938320 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9382975829999998 podStartE2EDuration="2.938297583s" podCreationTimestamp="2025-10-01 09:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:09:54.924507611 +0000 UTC m=+7451.766115106" watchObservedRunningTime="2025-10-01 09:09:54.938297583 +0000 UTC m=+7451.779905048" Oct 01 09:09:55 crc kubenswrapper[4837]: I1001 09:09:55.511956 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:09:55 crc kubenswrapper[4837]: I1001 09:09:55.618656 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:09:55 crc kubenswrapper[4837]: I1001 09:09:55.619062 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="dnsmasq-dns" containerID="cri-o://6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d" gracePeriod=10 Oct 01 09:09:55 crc kubenswrapper[4837]: E1001 09:09:55.810950 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadfd5762_5f62_4d1d_8e76_baa2f6ff7a91.slice/crio-conmon-6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadfd5762_5f62_4d1d_8e76_baa2f6ff7a91.slice/crio-6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:09:55 crc kubenswrapper[4837]: I1001 09:09:55.918577 4837 generic.go:334] "Generic (PLEG): container finished" podID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerID="6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d" exitCode=0 Oct 01 09:09:55 crc kubenswrapper[4837]: I1001 09:09:55.918947 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" event={"ID":"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91","Type":"ContainerDied","Data":"6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d"} Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.127034 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.211704 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrdcr\" (UniqueName: \"kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr\") pod \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.211786 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb\") pod \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.211838 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb\") pod \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.211926 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config\") pod \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.211958 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc\") pod \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\" (UID: \"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91\") " Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.221275 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr" (OuterVolumeSpecName: "kube-api-access-wrdcr") pod "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" (UID: "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91"). InnerVolumeSpecName "kube-api-access-wrdcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.269461 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" (UID: "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.271032 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" (UID: "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.276571 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" (UID: "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.280799 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config" (OuterVolumeSpecName: "config") pod "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" (UID: "adfd5762-5f62-4d1d-8e76-baa2f6ff7a91"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.314185 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrdcr\" (UniqueName: \"kubernetes.io/projected/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-kube-api-access-wrdcr\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.314216 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.314228 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.314238 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.314248 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.931503 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" event={"ID":"adfd5762-5f62-4d1d-8e76-baa2f6ff7a91","Type":"ContainerDied","Data":"395ea95b0e016e9bff142e8260e50644dcbc01bb9a58f7f022c8987425eb2ba5"} Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.931580 4837 scope.go:117] "RemoveContainer" containerID="6a4064a7dc28cfef4f16a1a44965d61bd6f3c1ea490fcb84186e28b8a29d862d" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.931815 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d649f7dc-rn6j5" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.960681 4837 scope.go:117] "RemoveContainer" containerID="4475635f91138bb3eb48b2e0c8530826a5717df6515d4e6a95d3a0de71fb8fb1" Oct 01 09:09:56 crc kubenswrapper[4837]: I1001 09:09:56.986404 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:09:57 crc kubenswrapper[4837]: I1001 09:09:57.000146 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9d649f7dc-rn6j5"] Oct 01 09:09:57 crc kubenswrapper[4837]: I1001 09:09:57.832807 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" path="/var/lib/kubelet/pods/adfd5762-5f62-4d1d-8e76-baa2f6ff7a91/volumes" Oct 01 09:10:03 crc kubenswrapper[4837]: I1001 09:10:03.247152 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:10:03 crc kubenswrapper[4837]: I1001 09:10:03.247777 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:10:04 crc kubenswrapper[4837]: I1001 09:10:04.258913 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.109:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:10:04 crc kubenswrapper[4837]: I1001 09:10:04.258949 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.109:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:10:13 crc kubenswrapper[4837]: I1001 09:10:13.255017 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:10:13 crc kubenswrapper[4837]: I1001 09:10:13.257275 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:10:13 crc kubenswrapper[4837]: I1001 09:10:13.263673 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:10:13 crc kubenswrapper[4837]: I1001 09:10:13.268987 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:10:14 crc kubenswrapper[4837]: I1001 09:10:14.131979 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:10:14 crc kubenswrapper[4837]: I1001 09:10:14.138610 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:10:23 crc kubenswrapper[4837]: I1001 09:10:23.079167 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:10:23 crc kubenswrapper[4837]: I1001 09:10:23.079760 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.050572 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:10:24 crc kubenswrapper[4837]: E1001 09:10:24.050967 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="dnsmasq-dns" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.050978 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="dnsmasq-dns" Oct 01 09:10:24 crc kubenswrapper[4837]: E1001 09:10:24.050985 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="init" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.050991 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="init" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.051167 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfd5762-5f62-4d1d-8e76-baa2f6ff7a91" containerName="dnsmasq-dns" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.052081 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.056900 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.058103 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-dc8ck" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.058258 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.058371 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.083825 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.101408 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.101705 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-log" containerID="cri-o://f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5" gracePeriod=30 Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.102222 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-httpd" containerID="cri-o://fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7" gracePeriod=30 Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.133028 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.133078 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.133119 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.133199 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.133220 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbrdb\" (UniqueName: \"kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.173778 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.174019 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-log" containerID="cri-o://b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d" gracePeriod=30 Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.174454 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-httpd" containerID="cri-o://7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5" gracePeriod=30 Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.197263 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.199263 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.207156 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.234407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.234446 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.234498 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.234602 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.234623 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbrdb\" (UniqueName: \"kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.235767 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.237156 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.237820 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.249634 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.260249 4837 generic.go:334] "Generic (PLEG): container finished" podID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerID="f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5" exitCode=143 Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.260296 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerDied","Data":"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5"} Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.261850 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbrdb\" (UniqueName: \"kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb\") pod \"horizon-9c654b6bf-zqlsr\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.336355 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.336438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7sln\" (UniqueName: \"kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.336648 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.336771 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.336872 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.382737 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.438334 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7sln\" (UniqueName: \"kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.438428 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.438486 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.438542 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.438610 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.439400 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.439664 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.440065 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.446143 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.454966 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7sln\" (UniqueName: \"kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln\") pod \"horizon-9d6c8cb89-mpv97\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.659728 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:24 crc kubenswrapper[4837]: I1001 09:10:24.810840 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:10:25 crc kubenswrapper[4837]: I1001 09:10:25.125152 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:10:25 crc kubenswrapper[4837]: W1001 09:10:25.138524 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod233df542_7d2d_4f35_98a6_779a325a06ff.slice/crio-aab2d7c6a75415dd09a011318d5a94339d231c89fc7f3ecdb0949dd09be5cf31 WatchSource:0}: Error finding container aab2d7c6a75415dd09a011318d5a94339d231c89fc7f3ecdb0949dd09be5cf31: Status 404 returned error can't find the container with id aab2d7c6a75415dd09a011318d5a94339d231c89fc7f3ecdb0949dd09be5cf31 Oct 01 09:10:25 crc kubenswrapper[4837]: I1001 09:10:25.277991 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerStarted","Data":"aab2d7c6a75415dd09a011318d5a94339d231c89fc7f3ecdb0949dd09be5cf31"} Oct 01 09:10:25 crc kubenswrapper[4837]: I1001 09:10:25.287445 4837 generic.go:334] "Generic (PLEG): container finished" podID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerID="b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d" exitCode=143 Oct 01 09:10:25 crc kubenswrapper[4837]: I1001 09:10:25.287534 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerDied","Data":"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d"} Oct 01 09:10:25 crc kubenswrapper[4837]: I1001 09:10:25.290578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerStarted","Data":"b9f07dd4158ae0790627ccbd43fe9851cfd354c791c436c7361f63e9fd0f7f7e"} Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.485741 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.523253 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.524815 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.528019 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.534902 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.578845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blkbx\" (UniqueName: \"kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.579457 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.579599 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.579724 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.579911 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.580179 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.580565 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.603389 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.628003 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.642781 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.668879 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695565 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695632 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695703 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blkbx\" (UniqueName: \"kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695744 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695821 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.695898 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.697267 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.698192 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.698935 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.702599 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.705484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.714193 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blkbx\" (UniqueName: \"kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.724098 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs\") pod \"horizon-748fbb56f4-dkm85\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.797016 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.797360 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvj4\" (UniqueName: \"kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.797612 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.797675 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.797909 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.798016 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.798105 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.877342 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.902844 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.902931 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903026 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903130 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903212 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvj4\" (UniqueName: \"kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903328 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.903778 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.904271 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.904517 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.907304 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.908869 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.917783 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.923344 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvj4\" (UniqueName: \"kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4\") pod \"horizon-7bdcd979b8-z6dmt\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:26 crc kubenswrapper[4837]: I1001 09:10:26.973242 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.345932 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.1.76:9292/healthcheck\": read tcp 10.217.0.2:42400->10.217.1.76:9292: read: connection reset by peer" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.346750 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.1.76:9292/healthcheck\": read tcp 10.217.0.2:42404->10.217.1.76:9292: read: connection reset by peer" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.379810 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:10:27 crc kubenswrapper[4837]: W1001 09:10:27.429128 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73b9ebde_027f_4ca0_8f2f_981b2f430fc3.slice/crio-75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d WatchSource:0}: Error finding container 75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d: Status 404 returned error can't find the container with id 75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.496110 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:10:27 crc kubenswrapper[4837]: W1001 09:10:27.501562 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod895440de_0fde_4148_9c31_4547b9c12764.slice/crio-25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316 WatchSource:0}: Error finding container 25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316: Status 404 returned error can't find the container with id 25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316 Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.699477 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.819799 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdrq4\" (UniqueName: \"kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.819876 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.819925 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.819981 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.820085 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.820190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.820232 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run\") pod \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\" (UID: \"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8\") " Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.821838 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs" (OuterVolumeSpecName: "logs") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.822265 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.826652 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts" (OuterVolumeSpecName: "scripts") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.833711 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4" (OuterVolumeSpecName: "kube-api-access-mdrq4") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "kube-api-access-mdrq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.872569 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.876504 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.888162 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data" (OuterVolumeSpecName: "config-data") pod "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" (UID: "0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.922959 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.922994 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.923003 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.923012 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdrq4\" (UniqueName: \"kubernetes.io/projected/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-kube-api-access-mdrq4\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.923022 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.923050 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.923061 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:27 crc kubenswrapper[4837]: I1001 09:10:27.967140 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023663 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023734 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023789 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023909 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023953 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.023980 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knlk2\" (UniqueName: \"kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2\") pod \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\" (UID: \"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a\") " Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.025459 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.025583 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs" (OuterVolumeSpecName: "logs") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.028661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2" (OuterVolumeSpecName: "kube-api-access-knlk2") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "kube-api-access-knlk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.051265 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts" (OuterVolumeSpecName: "scripts") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.057382 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.087198 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.098362 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data" (OuterVolumeSpecName: "config-data") pod "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" (UID: "0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127104 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127135 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127144 4837 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127152 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127160 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127169 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.127179 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knlk2\" (UniqueName: \"kubernetes.io/projected/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a-kube-api-access-knlk2\") on node \"crc\" DevicePath \"\"" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.335830 4837 generic.go:334] "Generic (PLEG): container finished" podID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerID="7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5" exitCode=0 Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.335892 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerDied","Data":"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.335909 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.335933 4837 scope.go:117] "RemoveContainer" containerID="7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.335922 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a","Type":"ContainerDied","Data":"9e63e420d74a667f78d1d315086d49f3651e72c6de8bfad1de1f707f0196bc9d"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.337178 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerStarted","Data":"25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.339125 4837 generic.go:334] "Generic (PLEG): container finished" podID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerID="fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7" exitCode=0 Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.339171 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.339187 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerDied","Data":"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.339216 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8","Type":"ContainerDied","Data":"0f7dd090eb480a56ecf28044b3bc51d595d76e2c0cb5587f3508a306c8af7c58"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.340632 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerStarted","Data":"75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d"} Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.382290 4837 scope.go:117] "RemoveContainer" containerID="b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.385772 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.398341 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.412062 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.420753 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.429340 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: E1001 09:10:28.429774 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.429792 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: E1001 09:10:28.429804 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.429810 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: E1001 09:10:28.429824 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.429829 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: E1001 09:10:28.429852 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.429858 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.430047 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.430069 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-httpd" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.430080 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.430093 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" containerName="glance-log" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.433777 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.435370 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-swd2f" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.435618 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.435722 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.435964 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.439582 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.466525 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.468100 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.471003 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.471192 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.489126 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533247 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-logs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533292 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533346 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfpkm\" (UniqueName: \"kubernetes.io/projected/50f83292-0b00-4281-8ed3-e072a08a6c5d-kube-api-access-pfpkm\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533433 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533461 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533483 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533512 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533531 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljcwn\" (UniqueName: \"kubernetes.io/projected/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-kube-api-access-ljcwn\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533561 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533608 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533642 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533661 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.533680 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636140 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636753 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljcwn\" (UniqueName: \"kubernetes.io/projected/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-kube-api-access-ljcwn\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636812 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636886 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636941 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.636969 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637000 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637041 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-logs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637063 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637099 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfpkm\" (UniqueName: \"kubernetes.io/projected/50f83292-0b00-4281-8ed3-e072a08a6c5d-kube-api-access-pfpkm\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637168 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637205 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637238 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.637951 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.638033 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50f83292-0b00-4281-8ed3-e072a08a6c5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.639533 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-logs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.640201 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.641891 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.641985 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.642252 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-scripts\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.642461 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.644374 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.644906 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.644991 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50f83292-0b00-4281-8ed3-e072a08a6c5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.646191 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-config-data\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.652101 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljcwn\" (UniqueName: \"kubernetes.io/projected/b3b9c094-9ba3-4e91-9ec3-168180bc56f0-kube-api-access-ljcwn\") pod \"glance-default-external-api-0\" (UID: \"b3b9c094-9ba3-4e91-9ec3-168180bc56f0\") " pod="openstack/glance-default-external-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.654057 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfpkm\" (UniqueName: \"kubernetes.io/projected/50f83292-0b00-4281-8ed3-e072a08a6c5d-kube-api-access-pfpkm\") pod \"glance-default-internal-api-0\" (UID: \"50f83292-0b00-4281-8ed3-e072a08a6c5d\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.758093 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:28 crc kubenswrapper[4837]: I1001 09:10:28.791855 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:10:29 crc kubenswrapper[4837]: I1001 09:10:29.842579 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8" path="/var/lib/kubelet/pods/0bdb56c8-a1bd-402b-89dd-4b0aacadf1a8/volumes" Oct 01 09:10:29 crc kubenswrapper[4837]: I1001 09:10:29.846141 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a" path="/var/lib/kubelet/pods/0cef128b-5dd5-4e78-b9b8-4dfd86fbf31a/volumes" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.397188 4837 scope.go:117] "RemoveContainer" containerID="7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5" Oct 01 09:10:33 crc kubenswrapper[4837]: E1001 09:10:33.397668 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5\": container with ID starting with 7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5 not found: ID does not exist" containerID="7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.397721 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5"} err="failed to get container status \"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5\": rpc error: code = NotFound desc = could not find container \"7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5\": container with ID starting with 7e1ef21168ca71eb0f265e777f741cf3136aec056aaa6c8b771e389e1d44e0c5 not found: ID does not exist" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.397743 4837 scope.go:117] "RemoveContainer" containerID="b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d" Oct 01 09:10:33 crc kubenswrapper[4837]: E1001 09:10:33.398026 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d\": container with ID starting with b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d not found: ID does not exist" containerID="b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.398057 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d"} err="failed to get container status \"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d\": rpc error: code = NotFound desc = could not find container \"b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d\": container with ID starting with b2a1e14e50d1244aa1f48541f33dbf5b7d62059f969fd488c7542fb58717dd5d not found: ID does not exist" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.398075 4837 scope.go:117] "RemoveContainer" containerID="fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.490515 4837 scope.go:117] "RemoveContainer" containerID="f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.682459 4837 scope.go:117] "RemoveContainer" containerID="fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7" Oct 01 09:10:33 crc kubenswrapper[4837]: E1001 09:10:33.683456 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7\": container with ID starting with fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7 not found: ID does not exist" containerID="fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.683515 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7"} err="failed to get container status \"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7\": rpc error: code = NotFound desc = could not find container \"fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7\": container with ID starting with fd56fb1a245283bf518137cab6a1c7974b982eb22c05ed3f669314c94d9f91b7 not found: ID does not exist" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.683538 4837 scope.go:117] "RemoveContainer" containerID="f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5" Oct 01 09:10:33 crc kubenswrapper[4837]: E1001 09:10:33.699356 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5\": container with ID starting with f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5 not found: ID does not exist" containerID="f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5" Oct 01 09:10:33 crc kubenswrapper[4837]: I1001 09:10:33.699406 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5"} err="failed to get container status \"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5\": rpc error: code = NotFound desc = could not find container \"f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5\": container with ID starting with f9fe97edac4e802e2618fad015b9a440e5870ad58b4a9e7731844fa944dd9ff5 not found: ID does not exist" Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.033896 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.127444 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:10:34 crc kubenswrapper[4837]: W1001 09:10:34.153990 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50f83292_0b00_4281_8ed3_e072a08a6c5d.slice/crio-c6e32f71551c94016144e61be8ad12b1c6fc16a49cf952a88e2f73878fb447bb WatchSource:0}: Error finding container c6e32f71551c94016144e61be8ad12b1c6fc16a49cf952a88e2f73878fb447bb: Status 404 returned error can't find the container with id c6e32f71551c94016144e61be8ad12b1c6fc16a49cf952a88e2f73878fb447bb Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.406369 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerStarted","Data":"2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.406412 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerStarted","Data":"ca479e7db22675145e10c8b899b5acb0398dcf5c5df96e323bc18dd8c6178aa9"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.406437 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9c654b6bf-zqlsr" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon-log" containerID="cri-o://ca479e7db22675145e10c8b899b5acb0398dcf5c5df96e323bc18dd8c6178aa9" gracePeriod=30 Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.406448 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9c654b6bf-zqlsr" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon" containerID="cri-o://2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45" gracePeriod=30 Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.409763 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerStarted","Data":"3341089752c8c88877dd6c27557df0e2b1fffeca21428dfe799f64441bbbaffe"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.409814 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerStarted","Data":"99fdcae14a1692b6ffa7e89638b0b14c7f8c0504163cdf0d012c71fb5f50a9f2"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.413979 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerStarted","Data":"a0d1e139373f3331c9de3b2a407fe939f1a63d2bf4690bc47f859170915dd769"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.414028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerStarted","Data":"6aa3e91be97793fe408de8982b33513f0bf73fa474a058382879348a8a4a77c5"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.416471 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50f83292-0b00-4281-8ed3-e072a08a6c5d","Type":"ContainerStarted","Data":"c6e32f71551c94016144e61be8ad12b1c6fc16a49cf952a88e2f73878fb447bb"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.418857 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b3b9c094-9ba3-4e91-9ec3-168180bc56f0","Type":"ContainerStarted","Data":"59b7bcc0457f90747847185cf7b0c2ce5ed827af440202fb4fd3fa224e9af1bf"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.422183 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerStarted","Data":"0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.422209 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerStarted","Data":"20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0"} Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.422338 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9d6c8cb89-mpv97" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon-log" containerID="cri-o://20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0" gracePeriod=30 Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.422483 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9d6c8cb89-mpv97" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon" containerID="cri-o://0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717" gracePeriod=30 Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.455334 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-9c654b6bf-zqlsr" podStartSLOduration=1.736386478 podStartE2EDuration="10.455316012s" podCreationTimestamp="2025-10-01 09:10:24 +0000 UTC" firstStartedPulling="2025-10-01 09:10:24.823073309 +0000 UTC m=+7481.664680764" lastFinishedPulling="2025-10-01 09:10:33.542002843 +0000 UTC m=+7490.383610298" observedRunningTime="2025-10-01 09:10:34.428588729 +0000 UTC m=+7491.270196194" watchObservedRunningTime="2025-10-01 09:10:34.455316012 +0000 UTC m=+7491.296923467" Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.456876 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bdcd979b8-z6dmt" podStartSLOduration=2.263537961 podStartE2EDuration="8.45687046s" podCreationTimestamp="2025-10-01 09:10:26 +0000 UTC" firstStartedPulling="2025-10-01 09:10:27.506074974 +0000 UTC m=+7484.347682439" lastFinishedPulling="2025-10-01 09:10:33.699407483 +0000 UTC m=+7490.541014938" observedRunningTime="2025-10-01 09:10:34.449509977 +0000 UTC m=+7491.291117432" watchObservedRunningTime="2025-10-01 09:10:34.45687046 +0000 UTC m=+7491.298477915" Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.484352 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-748fbb56f4-dkm85" podStartSLOduration=2.288476418 podStartE2EDuration="8.48433102s" podCreationTimestamp="2025-10-01 09:10:26 +0000 UTC" firstStartedPulling="2025-10-01 09:10:27.437672509 +0000 UTC m=+7484.279279964" lastFinishedPulling="2025-10-01 09:10:33.633527111 +0000 UTC m=+7490.475134566" observedRunningTime="2025-10-01 09:10:34.470929249 +0000 UTC m=+7491.312536704" watchObservedRunningTime="2025-10-01 09:10:34.48433102 +0000 UTC m=+7491.325938475" Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.497552 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-9d6c8cb89-mpv97" podStartSLOduration=2.005194808 podStartE2EDuration="10.497534768s" podCreationTimestamp="2025-10-01 09:10:24 +0000 UTC" firstStartedPulling="2025-10-01 09:10:25.141246162 +0000 UTC m=+7481.982853617" lastFinishedPulling="2025-10-01 09:10:33.633586122 +0000 UTC m=+7490.475193577" observedRunningTime="2025-10-01 09:10:34.488130484 +0000 UTC m=+7491.329737959" watchObservedRunningTime="2025-10-01 09:10:34.497534768 +0000 UTC m=+7491.339142223" Oct 01 09:10:34 crc kubenswrapper[4837]: I1001 09:10:34.660485 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.438423 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b3b9c094-9ba3-4e91-9ec3-168180bc56f0","Type":"ContainerStarted","Data":"04830b339278f1711cec5c1e985a1af9915eda3057604e1f2da7462e5048709e"} Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.438840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b3b9c094-9ba3-4e91-9ec3-168180bc56f0","Type":"ContainerStarted","Data":"5e5f0f3ffeb45610a4e6a5c4972513776abde18e1c42dc396510bf576745f870"} Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.441019 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50f83292-0b00-4281-8ed3-e072a08a6c5d","Type":"ContainerStarted","Data":"5f473d8b5ff2bbea86505b3551f9c740e53831952ea974b47a953a5255cd70f1"} Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.441054 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50f83292-0b00-4281-8ed3-e072a08a6c5d","Type":"ContainerStarted","Data":"c856d2bccb06320495efbfbd5841b731f120b7a121cc146272de4f0309fce8e0"} Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.474860 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.474840622 podStartE2EDuration="7.474840622s" podCreationTimestamp="2025-10-01 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:10:35.467774067 +0000 UTC m=+7492.309381532" watchObservedRunningTime="2025-10-01 09:10:35.474840622 +0000 UTC m=+7492.316448077" Oct 01 09:10:35 crc kubenswrapper[4837]: I1001 09:10:35.499828 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.49976442 podStartE2EDuration="7.49976442s" podCreationTimestamp="2025-10-01 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:10:35.493159026 +0000 UTC m=+7492.334766511" watchObservedRunningTime="2025-10-01 09:10:35.49976442 +0000 UTC m=+7492.341371885" Oct 01 09:10:36 crc kubenswrapper[4837]: I1001 09:10:36.877876 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:36 crc kubenswrapper[4837]: I1001 09:10:36.878295 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:36 crc kubenswrapper[4837]: I1001 09:10:36.973333 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:36 crc kubenswrapper[4837]: I1001 09:10:36.973403 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.760050 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.760535 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.792261 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.792735 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.802789 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.804131 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.835628 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:10:38 crc kubenswrapper[4837]: I1001 09:10:38.837459 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:10:39 crc kubenswrapper[4837]: I1001 09:10:39.485516 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:39 crc kubenswrapper[4837]: I1001 09:10:39.485726 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:39 crc kubenswrapper[4837]: I1001 09:10:39.485765 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:10:39 crc kubenswrapper[4837]: I1001 09:10:39.485776 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:10:41 crc kubenswrapper[4837]: I1001 09:10:41.853433 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:41 crc kubenswrapper[4837]: I1001 09:10:41.855422 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:10:42 crc kubenswrapper[4837]: I1001 09:10:42.034288 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:10:42 crc kubenswrapper[4837]: I1001 09:10:42.036639 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:10:44 crc kubenswrapper[4837]: I1001 09:10:44.383489 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:10:46 crc kubenswrapper[4837]: I1001 09:10:46.881319 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.112:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8443: connect: connection refused" Oct 01 09:10:46 crc kubenswrapper[4837]: I1001 09:10:46.976800 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.113:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8443: connect: connection refused" Oct 01 09:10:53 crc kubenswrapper[4837]: I1001 09:10:53.079204 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:10:53 crc kubenswrapper[4837]: I1001 09:10:53.079545 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:10:59 crc kubenswrapper[4837]: I1001 09:10:59.630121 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:10:59 crc kubenswrapper[4837]: I1001 09:10:59.647782 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:11:01 crc kubenswrapper[4837]: I1001 09:11:01.220566 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:11:01 crc kubenswrapper[4837]: I1001 09:11:01.355579 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:11:01 crc kubenswrapper[4837]: I1001 09:11:01.416590 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:11:01 crc kubenswrapper[4837]: I1001 09:11:01.744526 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon-log" containerID="cri-o://6aa3e91be97793fe408de8982b33513f0bf73fa474a058382879348a8a4a77c5" gracePeriod=30 Oct 01 09:11:01 crc kubenswrapper[4837]: I1001 09:11:01.744726 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" containerID="cri-o://a0d1e139373f3331c9de3b2a407fe939f1a63d2bf4690bc47f859170915dd769" gracePeriod=30 Oct 01 09:11:04 crc kubenswrapper[4837]: E1001 09:11:04.717549 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod233df542_7d2d_4f35_98a6_779a325a06ff.slice/crio-conmon-20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31812daf_db01_4188_9613_e9dbd16ce145.slice/crio-conmon-2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod233df542_7d2d_4f35_98a6_779a325a06ff.slice/crio-conmon-0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod233df542_7d2d_4f35_98a6_779a325a06ff.slice/crio-0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.779042 4837 generic.go:334] "Generic (PLEG): container finished" podID="233df542-7d2d-4f35-98a6-779a325a06ff" containerID="0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717" exitCode=137 Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.779085 4837 generic.go:334] "Generic (PLEG): container finished" podID="233df542-7d2d-4f35-98a6-779a325a06ff" containerID="20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0" exitCode=137 Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.779098 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerDied","Data":"0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717"} Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.779155 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerDied","Data":"20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0"} Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.782026 4837 generic.go:334] "Generic (PLEG): container finished" podID="31812daf-db01-4188-9613-e9dbd16ce145" containerID="2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45" exitCode=137 Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.782056 4837 generic.go:334] "Generic (PLEG): container finished" podID="31812daf-db01-4188-9613-e9dbd16ce145" containerID="ca479e7db22675145e10c8b899b5acb0398dcf5c5df96e323bc18dd8c6178aa9" exitCode=137 Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.782074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerDied","Data":"2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45"} Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.782096 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerDied","Data":"ca479e7db22675145e10c8b899b5acb0398dcf5c5df96e323bc18dd8c6178aa9"} Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.936890 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:11:04 crc kubenswrapper[4837]: I1001 09:11:04.941033 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.125777 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.125864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data\") pod \"31812daf-db01-4188-9613-e9dbd16ce145\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.125907 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs\") pod \"31812daf-db01-4188-9613-e9dbd16ce145\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.125982 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126019 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts\") pod \"31812daf-db01-4188-9613-e9dbd16ce145\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126620 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7sln\" (UniqueName: \"kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126706 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126747 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key\") pod \"31812daf-db01-4188-9613-e9dbd16ce145\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126782 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbrdb\" (UniqueName: \"kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb\") pod \"31812daf-db01-4188-9613-e9dbd16ce145\" (UID: \"31812daf-db01-4188-9613-e9dbd16ce145\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.126843 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.127348 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs" (OuterVolumeSpecName: "logs") pod "31812daf-db01-4188-9613-e9dbd16ce145" (UID: "31812daf-db01-4188-9613-e9dbd16ce145"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.127883 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs" (OuterVolumeSpecName: "logs") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.131973 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb" (OuterVolumeSpecName: "kube-api-access-lbrdb") pod "31812daf-db01-4188-9613-e9dbd16ce145" (UID: "31812daf-db01-4188-9613-e9dbd16ce145"). InnerVolumeSpecName "kube-api-access-lbrdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.131983 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "31812daf-db01-4188-9613-e9dbd16ce145" (UID: "31812daf-db01-4188-9613-e9dbd16ce145"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.133306 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.133320 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln" (OuterVolumeSpecName: "kube-api-access-m7sln") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff"). InnerVolumeSpecName "kube-api-access-m7sln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: E1001 09:11:05.152257 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts podName:233df542-7d2d-4f35-98a6-779a325a06ff nodeName:}" failed. No retries permitted until 2025-10-01 09:11:05.652227001 +0000 UTC m=+7522.493834466 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "scripts" (UniqueName: "kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff") : error deleting /var/lib/kubelet/pods/233df542-7d2d-4f35-98a6-779a325a06ff/volume-subpaths: remove /var/lib/kubelet/pods/233df542-7d2d-4f35-98a6-779a325a06ff/volume-subpaths: no such file or directory Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.152318 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts" (OuterVolumeSpecName: "scripts") pod "31812daf-db01-4188-9613-e9dbd16ce145" (UID: "31812daf-db01-4188-9613-e9dbd16ce145"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.152621 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data" (OuterVolumeSpecName: "config-data") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.165001 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data" (OuterVolumeSpecName: "config-data") pod "31812daf-db01-4188-9613-e9dbd16ce145" (UID: "31812daf-db01-4188-9613-e9dbd16ce145"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236042 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/233df542-7d2d-4f35-98a6-779a325a06ff-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236074 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31812daf-db01-4188-9613-e9dbd16ce145-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236096 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236107 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236116 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31812daf-db01-4188-9613-e9dbd16ce145-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236126 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7sln\" (UniqueName: \"kubernetes.io/projected/233df542-7d2d-4f35-98a6-779a325a06ff-kube-api-access-m7sln\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236141 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/233df542-7d2d-4f35-98a6-779a325a06ff-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236150 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31812daf-db01-4188-9613-e9dbd16ce145-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.236159 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbrdb\" (UniqueName: \"kubernetes.io/projected/31812daf-db01-4188-9613-e9dbd16ce145-kube-api-access-lbrdb\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.745465 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") pod \"233df542-7d2d-4f35-98a6-779a325a06ff\" (UID: \"233df542-7d2d-4f35-98a6-779a325a06ff\") " Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.745914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts" (OuterVolumeSpecName: "scripts") pod "233df542-7d2d-4f35-98a6-779a325a06ff" (UID: "233df542-7d2d-4f35-98a6-779a325a06ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.746197 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/233df542-7d2d-4f35-98a6-779a325a06ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.800160 4837 generic.go:334] "Generic (PLEG): container finished" podID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerID="a0d1e139373f3331c9de3b2a407fe939f1a63d2bf4690bc47f859170915dd769" exitCode=0 Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.800249 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerDied","Data":"a0d1e139373f3331c9de3b2a407fe939f1a63d2bf4690bc47f859170915dd769"} Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.803218 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9d6c8cb89-mpv97" event={"ID":"233df542-7d2d-4f35-98a6-779a325a06ff","Type":"ContainerDied","Data":"aab2d7c6a75415dd09a011318d5a94339d231c89fc7f3ecdb0949dd09be5cf31"} Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.803249 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9d6c8cb89-mpv97" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.803288 4837 scope.go:117] "RemoveContainer" containerID="0aeb377ca38c4a8d958314cb78adc035fb10372536a33f841db1e3b808b91717" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.807502 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c654b6bf-zqlsr" event={"ID":"31812daf-db01-4188-9613-e9dbd16ce145","Type":"ContainerDied","Data":"b9f07dd4158ae0790627ccbd43fe9851cfd354c791c436c7361f63e9fd0f7f7e"} Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.807602 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c654b6bf-zqlsr" Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.838983 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.845733 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9d6c8cb89-mpv97"] Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.862498 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.868411 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9c654b6bf-zqlsr"] Oct 01 09:11:05 crc kubenswrapper[4837]: I1001 09:11:05.988449 4837 scope.go:117] "RemoveContainer" containerID="20da8764fa9d53d53ca7b92f5da635fe467263ebf8235471c1370d33deac0aa0" Oct 01 09:11:06 crc kubenswrapper[4837]: I1001 09:11:06.010103 4837 scope.go:117] "RemoveContainer" containerID="2cd02c15d4daed97a27be8ce6a748f282f9b3fcf58ada0c15fc6465a48a42c45" Oct 01 09:11:06 crc kubenswrapper[4837]: I1001 09:11:06.042926 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-42cq8"] Oct 01 09:11:06 crc kubenswrapper[4837]: I1001 09:11:06.052761 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-42cq8"] Oct 01 09:11:06 crc kubenswrapper[4837]: I1001 09:11:06.204159 4837 scope.go:117] "RemoveContainer" containerID="ca479e7db22675145e10c8b899b5acb0398dcf5c5df96e323bc18dd8c6178aa9" Oct 01 09:11:06 crc kubenswrapper[4837]: I1001 09:11:06.879201 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.112:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8443: connect: connection refused" Oct 01 09:11:07 crc kubenswrapper[4837]: I1001 09:11:07.838643 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" path="/var/lib/kubelet/pods/233df542-7d2d-4f35-98a6-779a325a06ff/volumes" Oct 01 09:11:07 crc kubenswrapper[4837]: I1001 09:11:07.841298 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31812daf-db01-4188-9613-e9dbd16ce145" path="/var/lib/kubelet/pods/31812daf-db01-4188-9613-e9dbd16ce145/volumes" Oct 01 09:11:07 crc kubenswrapper[4837]: I1001 09:11:07.842920 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7892968b-bed5-4332-b510-12f2b12caefd" path="/var/lib/kubelet/pods/7892968b-bed5-4332-b510-12f2b12caefd/volumes" Oct 01 09:11:10 crc kubenswrapper[4837]: I1001 09:11:10.395036 4837 scope.go:117] "RemoveContainer" containerID="f842a1e3b920a2313fe5810dfec9ba48b40514c8c601dfe1edfe5c246785e96d" Oct 01 09:11:10 crc kubenswrapper[4837]: I1001 09:11:10.461009 4837 scope.go:117] "RemoveContainer" containerID="636dde0f86ee0b9dfdaa5155ee41599f956c39e647ffe136f2a8680a46555cea" Oct 01 09:11:10 crc kubenswrapper[4837]: I1001 09:11:10.495431 4837 scope.go:117] "RemoveContainer" containerID="8678c653aa5843282f890228a0a4127405fb6217649ad0916d4d7720c4db631d" Oct 01 09:11:16 crc kubenswrapper[4837]: I1001 09:11:16.049997 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-633e-account-create-hdn2v"] Oct 01 09:11:16 crc kubenswrapper[4837]: I1001 09:11:16.067441 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-633e-account-create-hdn2v"] Oct 01 09:11:16 crc kubenswrapper[4837]: I1001 09:11:16.879588 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.112:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8443: connect: connection refused" Oct 01 09:11:17 crc kubenswrapper[4837]: I1001 09:11:17.841082 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0196e178-95f4-4e0f-b09b-f420c71859bc" path="/var/lib/kubelet/pods/0196e178-95f4-4e0f-b09b-f420c71859bc/volumes" Oct 01 09:11:23 crc kubenswrapper[4837]: I1001 09:11:23.079365 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:11:23 crc kubenswrapper[4837]: I1001 09:11:23.080374 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:11:23 crc kubenswrapper[4837]: I1001 09:11:23.080449 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:11:23 crc kubenswrapper[4837]: I1001 09:11:23.081866 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:11:23 crc kubenswrapper[4837]: I1001 09:11:23.081936 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810" gracePeriod=600 Oct 01 09:11:24 crc kubenswrapper[4837]: I1001 09:11:24.029660 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810" exitCode=0 Oct 01 09:11:24 crc kubenswrapper[4837]: I1001 09:11:24.029759 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810"} Oct 01 09:11:24 crc kubenswrapper[4837]: I1001 09:11:24.030334 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5"} Oct 01 09:11:24 crc kubenswrapper[4837]: I1001 09:11:24.030375 4837 scope.go:117] "RemoveContainer" containerID="32753893305f5e95500edb966cf3a1148c2810eeaf2ca97bce804ed378c346d6" Oct 01 09:11:26 crc kubenswrapper[4837]: I1001 09:11:26.879057 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-748fbb56f4-dkm85" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.112:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8443: connect: connection refused" Oct 01 09:11:26 crc kubenswrapper[4837]: I1001 09:11:26.879932 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:11:31 crc kubenswrapper[4837]: I1001 09:11:31.043482 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-mbxg6"] Oct 01 09:11:31 crc kubenswrapper[4837]: I1001 09:11:31.058516 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-mbxg6"] Oct 01 09:11:31 crc kubenswrapper[4837]: I1001 09:11:31.831767 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef08c36-8f34-4174-bc6f-d6bdaa96a498" path="/var/lib/kubelet/pods/9ef08c36-8f34-4174-bc6f-d6bdaa96a498/volumes" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.124667 4837 generic.go:334] "Generic (PLEG): container finished" podID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerID="6aa3e91be97793fe408de8982b33513f0bf73fa474a058382879348a8a4a77c5" exitCode=137 Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.124898 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerDied","Data":"6aa3e91be97793fe408de8982b33513f0bf73fa474a058382879348a8a4a77c5"} Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.125253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-748fbb56f4-dkm85" event={"ID":"73b9ebde-027f-4ca0-8f2f-981b2f430fc3","Type":"ContainerDied","Data":"75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d"} Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.125275 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75abd7c393d1dec8139c8bb89d3bdbcabd467bebced208cb1849492b4943dd5d" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.178746 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.261578 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.262177 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs" (OuterVolumeSpecName: "logs") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.262214 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blkbx\" (UniqueName: \"kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.262273 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.262932 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.262959 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.263082 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.263109 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs\") pod \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\" (UID: \"73b9ebde-027f-4ca0-8f2f-981b2f430fc3\") " Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.263632 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.269144 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.270315 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx" (OuterVolumeSpecName: "kube-api-access-blkbx") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "kube-api-access-blkbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.292415 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data" (OuterVolumeSpecName: "config-data") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.296756 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts" (OuterVolumeSpecName: "scripts") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.303611 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.331842 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "73b9ebde-027f-4ca0-8f2f-981b2f430fc3" (UID: "73b9ebde-027f-4ca0-8f2f-981b2f430fc3"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371222 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blkbx\" (UniqueName: \"kubernetes.io/projected/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-kube-api-access-blkbx\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371266 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371280 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371294 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371307 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:32 crc kubenswrapper[4837]: I1001 09:11:32.371323 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73b9ebde-027f-4ca0-8f2f-981b2f430fc3-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:11:33 crc kubenswrapper[4837]: I1001 09:11:33.139301 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-748fbb56f4-dkm85" Oct 01 09:11:33 crc kubenswrapper[4837]: I1001 09:11:33.181988 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:11:33 crc kubenswrapper[4837]: I1001 09:11:33.208430 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-748fbb56f4-dkm85"] Oct 01 09:11:33 crc kubenswrapper[4837]: I1001 09:11:33.834796 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" path="/var/lib/kubelet/pods/73b9ebde-027f-4ca0-8f2f-981b2f430fc3/volumes" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.337582 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-575bccb888-ndbhd"] Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338393 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338405 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338424 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338429 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338451 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338457 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338467 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338472 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338481 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338487 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: E1001 09:12:05.338509 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338515 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338712 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338723 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338736 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338746 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b9ebde-027f-4ca0-8f2f-981b2f430fc3" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338763 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="233df542-7d2d-4f35-98a6-779a325a06ff" containerName="horizon" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.338773 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="31812daf-db01-4188-9613-e9dbd16ce145" containerName="horizon-log" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.339687 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.361165 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-575bccb888-ndbhd"] Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.441948 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-secret-key\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442056 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-tls-certs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442220 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-config-data\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-combined-ca-bundle\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-scripts\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442398 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df71e942-12b3-4d00-8238-bf93a99ab3a9-logs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.442487 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z98m8\" (UniqueName: \"kubernetes.io/projected/df71e942-12b3-4d00-8238-bf93a99ab3a9-kube-api-access-z98m8\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546266 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z98m8\" (UniqueName: \"kubernetes.io/projected/df71e942-12b3-4d00-8238-bf93a99ab3a9-kube-api-access-z98m8\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546412 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-secret-key\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546453 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-tls-certs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546517 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-config-data\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546541 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-combined-ca-bundle\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546559 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-scripts\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.546586 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df71e942-12b3-4d00-8238-bf93a99ab3a9-logs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.547406 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df71e942-12b3-4d00-8238-bf93a99ab3a9-logs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.548109 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-scripts\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.550147 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df71e942-12b3-4d00-8238-bf93a99ab3a9-config-data\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.553643 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-combined-ca-bundle\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.563374 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-tls-certs\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.564250 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df71e942-12b3-4d00-8238-bf93a99ab3a9-horizon-secret-key\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.574616 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z98m8\" (UniqueName: \"kubernetes.io/projected/df71e942-12b3-4d00-8238-bf93a99ab3a9-kube-api-access-z98m8\") pod \"horizon-575bccb888-ndbhd\" (UID: \"df71e942-12b3-4d00-8238-bf93a99ab3a9\") " pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:05 crc kubenswrapper[4837]: I1001 09:12:05.677972 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.146822 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-575bccb888-ndbhd"] Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.566516 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-575bccb888-ndbhd" event={"ID":"df71e942-12b3-4d00-8238-bf93a99ab3a9","Type":"ContainerStarted","Data":"44e5510612a746ce3c9923a780b9860ab388bacae53e5677379e4909af3e9fca"} Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.566588 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-575bccb888-ndbhd" event={"ID":"df71e942-12b3-4d00-8238-bf93a99ab3a9","Type":"ContainerStarted","Data":"4bc920925da8ebe070f232596d540d490c180393a1190e9cb0dd21d0149b58f1"} Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.566607 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-575bccb888-ndbhd" event={"ID":"df71e942-12b3-4d00-8238-bf93a99ab3a9","Type":"ContainerStarted","Data":"bd9ec0d74c4ff25df8303458df63b274d00be5d8492d2390bec90fc6c1f954f2"} Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.610522 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-575bccb888-ndbhd" podStartSLOduration=1.610500113 podStartE2EDuration="1.610500113s" podCreationTimestamp="2025-10-01 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:06.587827391 +0000 UTC m=+7583.429434846" watchObservedRunningTime="2025-10-01 09:12:06.610500113 +0000 UTC m=+7583.452107578" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.634037 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-fgb2r"] Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.637127 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.643804 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-fgb2r"] Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.666795 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rcvw\" (UniqueName: \"kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw\") pod \"heat-db-create-fgb2r\" (UID: \"33c79e5a-fb36-4447-a489-4a0493d1e56f\") " pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.769117 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rcvw\" (UniqueName: \"kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw\") pod \"heat-db-create-fgb2r\" (UID: \"33c79e5a-fb36-4447-a489-4a0493d1e56f\") " pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.788085 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rcvw\" (UniqueName: \"kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw\") pod \"heat-db-create-fgb2r\" (UID: \"33c79e5a-fb36-4447-a489-4a0493d1e56f\") " pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:06 crc kubenswrapper[4837]: I1001 09:12:06.962957 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:07 crc kubenswrapper[4837]: I1001 09:12:07.464634 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-fgb2r"] Oct 01 09:12:07 crc kubenswrapper[4837]: I1001 09:12:07.574848 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-fgb2r" event={"ID":"33c79e5a-fb36-4447-a489-4a0493d1e56f","Type":"ContainerStarted","Data":"e77a21b7349aebbf72bca3e51a98f2c9a692fc440ce8cc5394bcc7b80d565b67"} Oct 01 09:12:08 crc kubenswrapper[4837]: I1001 09:12:08.590737 4837 generic.go:334] "Generic (PLEG): container finished" podID="33c79e5a-fb36-4447-a489-4a0493d1e56f" containerID="574204b2ed6f85400d2db6410c3d21c66a0f961beb34b1c9994af10b8473d4a4" exitCode=0 Oct 01 09:12:08 crc kubenswrapper[4837]: I1001 09:12:08.590922 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-fgb2r" event={"ID":"33c79e5a-fb36-4447-a489-4a0493d1e56f","Type":"ContainerDied","Data":"574204b2ed6f85400d2db6410c3d21c66a0f961beb34b1c9994af10b8473d4a4"} Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.076715 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.145945 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rcvw\" (UniqueName: \"kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw\") pod \"33c79e5a-fb36-4447-a489-4a0493d1e56f\" (UID: \"33c79e5a-fb36-4447-a489-4a0493d1e56f\") " Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.151308 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw" (OuterVolumeSpecName: "kube-api-access-7rcvw") pod "33c79e5a-fb36-4447-a489-4a0493d1e56f" (UID: "33c79e5a-fb36-4447-a489-4a0493d1e56f"). InnerVolumeSpecName "kube-api-access-7rcvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.247986 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rcvw\" (UniqueName: \"kubernetes.io/projected/33c79e5a-fb36-4447-a489-4a0493d1e56f-kube-api-access-7rcvw\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.621187 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-fgb2r" event={"ID":"33c79e5a-fb36-4447-a489-4a0493d1e56f","Type":"ContainerDied","Data":"e77a21b7349aebbf72bca3e51a98f2c9a692fc440ce8cc5394bcc7b80d565b67"} Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.621257 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e77a21b7349aebbf72bca3e51a98f2c9a692fc440ce8cc5394bcc7b80d565b67" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.621359 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-fgb2r" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.693331 4837 scope.go:117] "RemoveContainer" containerID="e740f074ff84750272ebbf9ec6abeed13e389625d3cc27dc40991c639d9c9412" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.723407 4837 scope.go:117] "RemoveContainer" containerID="8f5e9afc1bf928299fb9aeacbeb044a6916edd586d2dee2febe5daba3bb7575b" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.765984 4837 scope.go:117] "RemoveContainer" containerID="a167d4e179f13cf7641b455adef11ceafbedcffb22ee26b1d271a980f92832ce" Oct 01 09:12:10 crc kubenswrapper[4837]: I1001 09:12:10.845495 4837 scope.go:117] "RemoveContainer" containerID="4fe01c92544f4d065eaea42022911abdba66f825dc77d9c6e1ba7233be6ed99b" Oct 01 09:12:15 crc kubenswrapper[4837]: I1001 09:12:15.678439 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:15 crc kubenswrapper[4837]: I1001 09:12:15.679178 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.741781 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-0856-account-create-5pgsg"] Oct 01 09:12:16 crc kubenswrapper[4837]: E1001 09:12:16.742428 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c79e5a-fb36-4447-a489-4a0493d1e56f" containerName="mariadb-database-create" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.742441 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c79e5a-fb36-4447-a489-4a0493d1e56f" containerName="mariadb-database-create" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.742674 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="33c79e5a-fb36-4447-a489-4a0493d1e56f" containerName="mariadb-database-create" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.743311 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.745680 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.754771 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-0856-account-create-5pgsg"] Oct 01 09:12:16 crc kubenswrapper[4837]: I1001 09:12:16.903185 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrnd\" (UniqueName: \"kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd\") pod \"heat-0856-account-create-5pgsg\" (UID: \"fc46596a-3455-475b-8f07-4c66ac27b9c6\") " pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.004714 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrnd\" (UniqueName: \"kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd\") pod \"heat-0856-account-create-5pgsg\" (UID: \"fc46596a-3455-475b-8f07-4c66ac27b9c6\") " pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.022453 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrnd\" (UniqueName: \"kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd\") pod \"heat-0856-account-create-5pgsg\" (UID: \"fc46596a-3455-475b-8f07-4c66ac27b9c6\") " pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.051030 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-dl2w4"] Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.059763 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-dl2w4"] Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.066263 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.572328 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-0856-account-create-5pgsg"] Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.714733 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0856-account-create-5pgsg" event={"ID":"fc46596a-3455-475b-8f07-4c66ac27b9c6","Type":"ContainerStarted","Data":"f123fbcd005ffc6f4348cde04ee84d04d08947e8ef3f428e3f4d89ad656eaab6"} Oct 01 09:12:17 crc kubenswrapper[4837]: I1001 09:12:17.838280 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcaeed1f-08d9-4633-bc53-c0f2cb10f38c" path="/var/lib/kubelet/pods/bcaeed1f-08d9-4633-bc53-c0f2cb10f38c/volumes" Oct 01 09:12:18 crc kubenswrapper[4837]: I1001 09:12:18.732322 4837 generic.go:334] "Generic (PLEG): container finished" podID="fc46596a-3455-475b-8f07-4c66ac27b9c6" containerID="2640c611abcb31747b8eaf066fa23193072bbccfd334f1e7b791ded21f1a5aa7" exitCode=0 Oct 01 09:12:18 crc kubenswrapper[4837]: I1001 09:12:18.732384 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0856-account-create-5pgsg" event={"ID":"fc46596a-3455-475b-8f07-4c66ac27b9c6","Type":"ContainerDied","Data":"2640c611abcb31747b8eaf066fa23193072bbccfd334f1e7b791ded21f1a5aa7"} Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.139147 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.290630 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqrnd\" (UniqueName: \"kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd\") pod \"fc46596a-3455-475b-8f07-4c66ac27b9c6\" (UID: \"fc46596a-3455-475b-8f07-4c66ac27b9c6\") " Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.298046 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd" (OuterVolumeSpecName: "kube-api-access-vqrnd") pod "fc46596a-3455-475b-8f07-4c66ac27b9c6" (UID: "fc46596a-3455-475b-8f07-4c66ac27b9c6"). InnerVolumeSpecName "kube-api-access-vqrnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.393734 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqrnd\" (UniqueName: \"kubernetes.io/projected/fc46596a-3455-475b-8f07-4c66ac27b9c6-kube-api-access-vqrnd\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.752938 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0856-account-create-5pgsg" event={"ID":"fc46596a-3455-475b-8f07-4c66ac27b9c6","Type":"ContainerDied","Data":"f123fbcd005ffc6f4348cde04ee84d04d08947e8ef3f428e3f4d89ad656eaab6"} Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.752975 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f123fbcd005ffc6f4348cde04ee84d04d08947e8ef3f428e3f4d89ad656eaab6" Oct 01 09:12:20 crc kubenswrapper[4837]: I1001 09:12:20.753023 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0856-account-create-5pgsg" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.842726 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-56ckh"] Oct 01 09:12:21 crc kubenswrapper[4837]: E1001 09:12:21.844486 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc46596a-3455-475b-8f07-4c66ac27b9c6" containerName="mariadb-account-create" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.844636 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc46596a-3455-475b-8f07-4c66ac27b9c6" containerName="mariadb-account-create" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.845160 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc46596a-3455-475b-8f07-4c66ac27b9c6" containerName="mariadb-account-create" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.846249 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-56ckh"] Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.846553 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.851990 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2stsz" Oct 01 09:12:21 crc kubenswrapper[4837]: I1001 09:12:21.853093 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.029756 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfc5r\" (UniqueName: \"kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.029857 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.029979 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.131971 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.132428 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.132630 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfc5r\" (UniqueName: \"kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.137996 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.138613 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.152538 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfc5r\" (UniqueName: \"kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r\") pod \"heat-db-sync-56ckh\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.174264 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.699549 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-56ckh"] Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.714036 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:12:22 crc kubenswrapper[4837]: I1001 09:12:22.776525 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-56ckh" event={"ID":"d5561324-9c39-4292-80a9-cad1343b9d87","Type":"ContainerStarted","Data":"6b23ff960b667e340074e7ba386a959ddb6d0af7ea1646770ace6e78241c569b"} Oct 01 09:12:27 crc kubenswrapper[4837]: I1001 09:12:27.039542 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b796-account-create-7g2r6"] Oct 01 09:12:27 crc kubenswrapper[4837]: I1001 09:12:27.052732 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b796-account-create-7g2r6"] Oct 01 09:12:27 crc kubenswrapper[4837]: I1001 09:12:27.356467 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:27 crc kubenswrapper[4837]: I1001 09:12:27.847977 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="052fcef6-6d55-4cb1-9b9a-111a49189b31" path="/var/lib/kubelet/pods/052fcef6-6d55-4cb1-9b9a-111a49189b31/volumes" Oct 01 09:12:28 crc kubenswrapper[4837]: I1001 09:12:28.870194 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-575bccb888-ndbhd" Oct 01 09:12:28 crc kubenswrapper[4837]: I1001 09:12:28.944686 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:12:28 crc kubenswrapper[4837]: I1001 09:12:28.944999 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon-log" containerID="cri-o://99fdcae14a1692b6ffa7e89638b0b14c7f8c0504163cdf0d012c71fb5f50a9f2" gracePeriod=30 Oct 01 09:12:28 crc kubenswrapper[4837]: I1001 09:12:28.945409 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" containerID="cri-o://3341089752c8c88877dd6c27557df0e2b1fffeca21428dfe799f64441bbbaffe" gracePeriod=30 Oct 01 09:12:31 crc kubenswrapper[4837]: I1001 09:12:31.897022 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-56ckh" event={"ID":"d5561324-9c39-4292-80a9-cad1343b9d87","Type":"ContainerStarted","Data":"58c6b92f330f632a64066c14383d9ce1b675787311bef343511c77e7df449a93"} Oct 01 09:12:31 crc kubenswrapper[4837]: I1001 09:12:31.919489 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-56ckh" podStartSLOduration=2.8116499839999998 podStartE2EDuration="10.919472128s" podCreationTimestamp="2025-10-01 09:12:21 +0000 UTC" firstStartedPulling="2025-10-01 09:12:22.713754313 +0000 UTC m=+7599.555361768" lastFinishedPulling="2025-10-01 09:12:30.821576427 +0000 UTC m=+7607.663183912" observedRunningTime="2025-10-01 09:12:31.917980762 +0000 UTC m=+7608.759588227" watchObservedRunningTime="2025-10-01 09:12:31.919472128 +0000 UTC m=+7608.761079593" Oct 01 09:12:32 crc kubenswrapper[4837]: I1001 09:12:32.941721 4837 generic.go:334] "Generic (PLEG): container finished" podID="895440de-0fde-4148-9c31-4547b9c12764" containerID="3341089752c8c88877dd6c27557df0e2b1fffeca21428dfe799f64441bbbaffe" exitCode=0 Oct 01 09:12:32 crc kubenswrapper[4837]: I1001 09:12:32.942075 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerDied","Data":"3341089752c8c88877dd6c27557df0e2b1fffeca21428dfe799f64441bbbaffe"} Oct 01 09:12:33 crc kubenswrapper[4837]: I1001 09:12:33.960505 4837 generic.go:334] "Generic (PLEG): container finished" podID="d5561324-9c39-4292-80a9-cad1343b9d87" containerID="58c6b92f330f632a64066c14383d9ce1b675787311bef343511c77e7df449a93" exitCode=0 Oct 01 09:12:33 crc kubenswrapper[4837]: I1001 09:12:33.960567 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-56ckh" event={"ID":"d5561324-9c39-4292-80a9-cad1343b9d87","Type":"ContainerDied","Data":"58c6b92f330f632a64066c14383d9ce1b675787311bef343511c77e7df449a93"} Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.377374 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.520589 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data\") pod \"d5561324-9c39-4292-80a9-cad1343b9d87\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.520625 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle\") pod \"d5561324-9c39-4292-80a9-cad1343b9d87\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.520649 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfc5r\" (UniqueName: \"kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r\") pod \"d5561324-9c39-4292-80a9-cad1343b9d87\" (UID: \"d5561324-9c39-4292-80a9-cad1343b9d87\") " Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.527087 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r" (OuterVolumeSpecName: "kube-api-access-sfc5r") pod "d5561324-9c39-4292-80a9-cad1343b9d87" (UID: "d5561324-9c39-4292-80a9-cad1343b9d87"). InnerVolumeSpecName "kube-api-access-sfc5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.574539 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5561324-9c39-4292-80a9-cad1343b9d87" (UID: "d5561324-9c39-4292-80a9-cad1343b9d87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.594276 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data" (OuterVolumeSpecName: "config-data") pod "d5561324-9c39-4292-80a9-cad1343b9d87" (UID: "d5561324-9c39-4292-80a9-cad1343b9d87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.623730 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.623786 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5561324-9c39-4292-80a9-cad1343b9d87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.623801 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfc5r\" (UniqueName: \"kubernetes.io/projected/d5561324-9c39-4292-80a9-cad1343b9d87-kube-api-access-sfc5r\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.984491 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-56ckh" event={"ID":"d5561324-9c39-4292-80a9-cad1343b9d87","Type":"ContainerDied","Data":"6b23ff960b667e340074e7ba386a959ddb6d0af7ea1646770ace6e78241c569b"} Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.984532 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b23ff960b667e340074e7ba386a959ddb6d0af7ea1646770ace6e78241c569b" Oct 01 09:12:35 crc kubenswrapper[4837]: I1001 09:12:35.984600 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-56ckh" Oct 01 09:12:36 crc kubenswrapper[4837]: I1001 09:12:36.975625 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.113:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8443: connect: connection refused" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.188793 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:12:37 crc kubenswrapper[4837]: E1001 09:12:37.189229 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5561324-9c39-4292-80a9-cad1343b9d87" containerName="heat-db-sync" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.189241 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5561324-9c39-4292-80a9-cad1343b9d87" containerName="heat-db-sync" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.189405 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5561324-9c39-4292-80a9-cad1343b9d87" containerName="heat-db-sync" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.190618 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.195482 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.195647 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-2stsz" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.201866 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.219543 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.228674 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.230021 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.236299 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.256143 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.269277 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrng2\" (UniqueName: \"kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.269337 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.269429 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.269474 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.296742 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.298172 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.302268 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.323928 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.370925 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371008 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371136 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371158 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb6gg\" (UniqueName: \"kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371217 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371287 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrng2\" (UniqueName: \"kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.371355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.379491 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.386406 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.388382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.390070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrng2\" (UniqueName: \"kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2\") pod \"heat-cfnapi-8c5bdbcc6-znf2l\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473230 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473293 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473360 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473384 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473415 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb6gg\" (UniqueName: \"kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473448 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473547 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb54x\" (UniqueName: \"kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.473604 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.479452 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.479863 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.480256 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.489247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb6gg\" (UniqueName: \"kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg\") pod \"heat-engine-678489d5c5-5rs4j\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.524514 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.559785 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.575180 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.575232 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.575283 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb54x\" (UniqueName: \"kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.575302 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.579189 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.582305 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.590848 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb54x\" (UniqueName: \"kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.591437 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom\") pod \"heat-api-7f88999dcd-56w72\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:37 crc kubenswrapper[4837]: I1001 09:12:37.624116 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:38 crc kubenswrapper[4837]: I1001 09:12:38.010246 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:12:38 crc kubenswrapper[4837]: W1001 09:12:38.011717 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc685b9c0_c2df_4408_8fd9_9bebdf59b732.slice/crio-ff94b1880ffa3894afe855b6baaab988581015dace4fa9bcc3daca8b0f6881e7 WatchSource:0}: Error finding container ff94b1880ffa3894afe855b6baaab988581015dace4fa9bcc3daca8b0f6881e7: Status 404 returned error can't find the container with id ff94b1880ffa3894afe855b6baaab988581015dace4fa9bcc3daca8b0f6881e7 Oct 01 09:12:38 crc kubenswrapper[4837]: W1001 09:12:38.097494 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3446a7a3_bb73_48be_ad29_734880b2426e.slice/crio-63a6f5d253b093993b0a1e1e7997f767208022fdce9623fb2979d4975ce9a5cd WatchSource:0}: Error finding container 63a6f5d253b093993b0a1e1e7997f767208022fdce9623fb2979d4975ce9a5cd: Status 404 returned error can't find the container with id 63a6f5d253b093993b0a1e1e7997f767208022fdce9623fb2979d4975ce9a5cd Oct 01 09:12:38 crc kubenswrapper[4837]: I1001 09:12:38.099609 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:38 crc kubenswrapper[4837]: I1001 09:12:38.196044 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.021824 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" event={"ID":"3446a7a3-bb73-48be-ad29-734880b2426e","Type":"ContainerStarted","Data":"63a6f5d253b093993b0a1e1e7997f767208022fdce9623fb2979d4975ce9a5cd"} Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.028315 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f88999dcd-56w72" event={"ID":"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b","Type":"ContainerStarted","Data":"ddc15214c960b8dbb4c8f0764c1c3ab20303404744ba1fdc55ccfbe53cda4153"} Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.037180 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678489d5c5-5rs4j" event={"ID":"c685b9c0-c2df-4408-8fd9-9bebdf59b732","Type":"ContainerStarted","Data":"8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f"} Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.037223 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678489d5c5-5rs4j" event={"ID":"c685b9c0-c2df-4408-8fd9-9bebdf59b732","Type":"ContainerStarted","Data":"ff94b1880ffa3894afe855b6baaab988581015dace4fa9bcc3daca8b0f6881e7"} Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.038255 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.049905 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wlkmb"] Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.058175 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wlkmb"] Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.074396 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-678489d5c5-5rs4j" podStartSLOduration=2.074377661 podStartE2EDuration="2.074377661s" podCreationTimestamp="2025-10-01 09:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:39.060013146 +0000 UTC m=+7615.901620611" watchObservedRunningTime="2025-10-01 09:12:39.074377661 +0000 UTC m=+7615.915985116" Oct 01 09:12:39 crc kubenswrapper[4837]: I1001 09:12:39.826143 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36fcdeac-e848-4460-93e2-ea6173928012" path="/var/lib/kubelet/pods/36fcdeac-e848-4460-93e2-ea6173928012/volumes" Oct 01 09:12:40 crc kubenswrapper[4837]: I1001 09:12:40.047203 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" event={"ID":"3446a7a3-bb73-48be-ad29-734880b2426e","Type":"ContainerStarted","Data":"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a"} Oct 01 09:12:40 crc kubenswrapper[4837]: I1001 09:12:40.047576 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:40 crc kubenswrapper[4837]: I1001 09:12:40.049401 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f88999dcd-56w72" event={"ID":"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b","Type":"ContainerStarted","Data":"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a"} Oct 01 09:12:40 crc kubenswrapper[4837]: I1001 09:12:40.070532 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" podStartSLOduration=1.500237856 podStartE2EDuration="3.070513933s" podCreationTimestamp="2025-10-01 09:12:37 +0000 UTC" firstStartedPulling="2025-10-01 09:12:38.100149253 +0000 UTC m=+7614.941756708" lastFinishedPulling="2025-10-01 09:12:39.67042534 +0000 UTC m=+7616.512032785" observedRunningTime="2025-10-01 09:12:40.06114796 +0000 UTC m=+7616.902755415" watchObservedRunningTime="2025-10-01 09:12:40.070513933 +0000 UTC m=+7616.912121388" Oct 01 09:12:40 crc kubenswrapper[4837]: I1001 09:12:40.078083 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7f88999dcd-56w72" podStartSLOduration=1.605625906 podStartE2EDuration="3.078067089s" podCreationTimestamp="2025-10-01 09:12:37 +0000 UTC" firstStartedPulling="2025-10-01 09:12:38.196444419 +0000 UTC m=+7615.038051894" lastFinishedPulling="2025-10-01 09:12:39.668885622 +0000 UTC m=+7616.510493077" observedRunningTime="2025-10-01 09:12:40.076401538 +0000 UTC m=+7616.918008993" watchObservedRunningTime="2025-10-01 09:12:40.078067089 +0000 UTC m=+7616.919674544" Oct 01 09:12:41 crc kubenswrapper[4837]: I1001 09:12:41.063634 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.616664 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-b8cc69c8f-zc64n"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.618521 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.628105 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b8cc69c8f-zc64n"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.706503 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8sfk\" (UniqueName: \"kubernetes.io/projected/d9393af6-558d-4e75-8602-f9e8d09ceb7f-kube-api-access-g8sfk\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.706829 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-combined-ca-bundle\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.706875 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data-custom\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.706913 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.717226 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.718586 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.737648 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.741522 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.755081 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.765495 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808413 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808463 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808501 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lz2j\" (UniqueName: \"kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808643 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808713 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l5j8\" (UniqueName: \"kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808748 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8sfk\" (UniqueName: \"kubernetes.io/projected/d9393af6-558d-4e75-8602-f9e8d09ceb7f-kube-api-access-g8sfk\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808771 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808866 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.808911 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.809020 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-combined-ca-bundle\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.809286 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data-custom\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.809381 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.821467 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-combined-ca-bundle\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.822188 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data-custom\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.822350 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9393af6-558d-4e75-8602-f9e8d09ceb7f-config-data\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.826005 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8sfk\" (UniqueName: \"kubernetes.io/projected/d9393af6-558d-4e75-8602-f9e8d09ceb7f-kube-api-access-g8sfk\") pod \"heat-engine-b8cc69c8f-zc64n\" (UID: \"d9393af6-558d-4e75-8602-f9e8d09ceb7f\") " pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.911439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.911952 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912072 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912098 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912127 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lz2j\" (UniqueName: \"kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912148 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912168 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l5j8\" (UniqueName: \"kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.912185 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.915823 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.921318 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.921559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.921629 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.921799 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.922673 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.931042 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l5j8\" (UniqueName: \"kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8\") pod \"heat-cfnapi-844c4df74c-t57kg\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.940383 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:43 crc kubenswrapper[4837]: I1001 09:12:43.954423 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lz2j\" (UniqueName: \"kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j\") pod \"heat-api-566cd58fb5-4lrgf\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.034536 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.060968 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.453217 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b8cc69c8f-zc64n"] Oct 01 09:12:44 crc kubenswrapper[4837]: W1001 09:12:44.455280 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9393af6_558d_4e75_8602_f9e8d09ceb7f.slice/crio-c79711b3caf2c9430332c92c6525e3ab02217c9dd808ff1d08debb4aa7cdab8b WatchSource:0}: Error finding container c79711b3caf2c9430332c92c6525e3ab02217c9dd808ff1d08debb4aa7cdab8b: Status 404 returned error can't find the container with id c79711b3caf2c9430332c92c6525e3ab02217c9dd808ff1d08debb4aa7cdab8b Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.616597 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.626973 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:44 crc kubenswrapper[4837]: W1001 09:12:44.627038 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75d8189f_c5f0_4328_b7ca_c979882a0b97.slice/crio-08ad549f9800a176cfdb7407fb2ba9aa6047b1ccfdedfabb69149c12b6976496 WatchSource:0}: Error finding container 08ad549f9800a176cfdb7407fb2ba9aa6047b1ccfdedfabb69149c12b6976496: Status 404 returned error can't find the container with id 08ad549f9800a176cfdb7407fb2ba9aa6047b1ccfdedfabb69149c12b6976496 Oct 01 09:12:44 crc kubenswrapper[4837]: W1001 09:12:44.630794 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f236154_f3c6_4347_aaa1_f0c338c4864d.slice/crio-7f6a38106a076d4e70173bc22ee07ca192376d51e752d7cba984832a762d5496 WatchSource:0}: Error finding container 7f6a38106a076d4e70173bc22ee07ca192376d51e752d7cba984832a762d5496: Status 404 returned error can't find the container with id 7f6a38106a076d4e70173bc22ee07ca192376d51e752d7cba984832a762d5496 Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.751577 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.751767 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-7f88999dcd-56w72" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" containerID="cri-o://d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a" gracePeriod=60 Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.763981 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-7f88999dcd-56w72" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.122:8004/healthcheck\": EOF" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.781738 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5dc5f6c468-wpmkd"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.784410 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.786894 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.786936 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.796427 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc5f6c468-wpmkd"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.820612 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.821764 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" containerID="cri-o://01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a" gracePeriod=60 Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829318 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829403 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data-custom\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829439 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-combined-ca-bundle\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829457 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-public-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829476 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-internal-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.829527 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5sxf\" (UniqueName: \"kubernetes.io/projected/da4dea0d-4b10-456a-9704-a6012b34bd40-kube-api-access-k5sxf\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.834833 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.121:8000/healthcheck\": EOF" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.840212 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-77ffb6c584-qdkpl"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.841459 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.850109 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.850328 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.863289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77ffb6c584-qdkpl"] Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931014 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-public-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931060 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931091 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931227 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data-custom\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931255 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-combined-ca-bundle\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931292 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data-custom\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-combined-ca-bundle\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931373 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-public-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931396 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-internal-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931435 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-internal-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931504 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwg77\" (UniqueName: \"kubernetes.io/projected/6c720536-3f20-4c58-b914-ddbd3c52645b-kube-api-access-gwg77\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.931532 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5sxf\" (UniqueName: \"kubernetes.io/projected/da4dea0d-4b10-456a-9704-a6012b34bd40-kube-api-access-k5sxf\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.938070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-combined-ca-bundle\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.944823 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data-custom\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.947167 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-public-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.947280 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-internal-tls-certs\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.948186 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da4dea0d-4b10-456a-9704-a6012b34bd40-config-data\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:44 crc kubenswrapper[4837]: I1001 09:12:44.963895 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5sxf\" (UniqueName: \"kubernetes.io/projected/da4dea0d-4b10-456a-9704-a6012b34bd40-kube-api-access-k5sxf\") pod \"heat-api-5dc5f6c468-wpmkd\" (UID: \"da4dea0d-4b10-456a-9704-a6012b34bd40\") " pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034011 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-public-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034361 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data-custom\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034643 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-combined-ca-bundle\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034833 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-internal-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.034965 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwg77\" (UniqueName: \"kubernetes.io/projected/6c720536-3f20-4c58-b914-ddbd3c52645b-kube-api-access-gwg77\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.040479 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-public-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.041326 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.042070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-combined-ca-bundle\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.043407 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-config-data-custom\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.049189 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c720536-3f20-4c58-b914-ddbd3c52645b-internal-tls-certs\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.052268 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwg77\" (UniqueName: \"kubernetes.io/projected/6c720536-3f20-4c58-b914-ddbd3c52645b-kube-api-access-gwg77\") pod \"heat-cfnapi-77ffb6c584-qdkpl\" (UID: \"6c720536-3f20-4c58-b914-ddbd3c52645b\") " pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.099815 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.103638 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b8cc69c8f-zc64n" event={"ID":"d9393af6-558d-4e75-8602-f9e8d09ceb7f","Type":"ContainerStarted","Data":"e528bbdc89477e9bdaa46ce2a1ba5610bf6beb202845a601f3b437c37be53da7"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.103714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b8cc69c8f-zc64n" event={"ID":"d9393af6-558d-4e75-8602-f9e8d09ceb7f","Type":"ContainerStarted","Data":"c79711b3caf2c9430332c92c6525e3ab02217c9dd808ff1d08debb4aa7cdab8b"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.105489 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566cd58fb5-4lrgf" event={"ID":"75d8189f-c5f0-4328-b7ca-c979882a0b97","Type":"ContainerStarted","Data":"ce019f2b04dbaefe31b414be4203d9617d29bbb19556b4aef9311ab4c9d18e48"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.105619 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566cd58fb5-4lrgf" event={"ID":"75d8189f-c5f0-4328-b7ca-c979882a0b97","Type":"ContainerStarted","Data":"08ad549f9800a176cfdb7407fb2ba9aa6047b1ccfdedfabb69149c12b6976496"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.106237 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.107886 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-844c4df74c-t57kg" event={"ID":"3f236154-f3c6-4347-aaa1-f0c338c4864d","Type":"ContainerStarted","Data":"02f2d2c9a6bcb5aa3edee55129c8c41bcf6839c78b9645afd18010e96ab3db9c"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.107985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-844c4df74c-t57kg" event={"ID":"3f236154-f3c6-4347-aaa1-f0c338c4864d","Type":"ContainerStarted","Data":"7f6a38106a076d4e70173bc22ee07ca192376d51e752d7cba984832a762d5496"} Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.108078 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.138772 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-b8cc69c8f-zc64n" podStartSLOduration=2.138750434 podStartE2EDuration="2.138750434s" podCreationTimestamp="2025-10-01 09:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:45.122255447 +0000 UTC m=+7621.963862902" watchObservedRunningTime="2025-10-01 09:12:45.138750434 +0000 UTC m=+7621.980357889" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.151544 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-566cd58fb5-4lrgf" podStartSLOduration=2.151525382 podStartE2EDuration="2.151525382s" podCreationTimestamp="2025-10-01 09:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:45.137118514 +0000 UTC m=+7621.978725969" watchObservedRunningTime="2025-10-01 09:12:45.151525382 +0000 UTC m=+7621.993132837" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.157492 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-844c4df74c-t57kg" podStartSLOduration=2.157476589 podStartE2EDuration="2.157476589s" podCreationTimestamp="2025-10-01 09:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:45.152989208 +0000 UTC m=+7621.994596663" watchObservedRunningTime="2025-10-01 09:12:45.157476589 +0000 UTC m=+7621.999084044" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.237567 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.578352 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc5f6c468-wpmkd"] Oct 01 09:12:45 crc kubenswrapper[4837]: W1001 09:12:45.581323 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda4dea0d_4b10_456a_9704_a6012b34bd40.slice/crio-47299c1505a1e034598329ed24639037d5ce98aa2f0915bd8a6ef62786fa7614 WatchSource:0}: Error finding container 47299c1505a1e034598329ed24639037d5ce98aa2f0915bd8a6ef62786fa7614: Status 404 returned error can't find the container with id 47299c1505a1e034598329ed24639037d5ce98aa2f0915bd8a6ef62786fa7614 Oct 01 09:12:45 crc kubenswrapper[4837]: I1001 09:12:45.758154 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77ffb6c584-qdkpl"] Oct 01 09:12:45 crc kubenswrapper[4837]: W1001 09:12:45.764800 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c720536_3f20_4c58_b914_ddbd3c52645b.slice/crio-61d69e2408bafa4a9edb18a50b43117648429d3868f7ed9479b36fdb14af99c3 WatchSource:0}: Error finding container 61d69e2408bafa4a9edb18a50b43117648429d3868f7ed9479b36fdb14af99c3: Status 404 returned error can't find the container with id 61d69e2408bafa4a9edb18a50b43117648429d3868f7ed9479b36fdb14af99c3 Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.119837 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc5f6c468-wpmkd" event={"ID":"da4dea0d-4b10-456a-9704-a6012b34bd40","Type":"ContainerStarted","Data":"ad5f6c03cdd519d9a60f6bdb1f890af08c48585122f66d7d0211fc300741af90"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.120255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc5f6c468-wpmkd" event={"ID":"da4dea0d-4b10-456a-9704-a6012b34bd40","Type":"ContainerStarted","Data":"47299c1505a1e034598329ed24639037d5ce98aa2f0915bd8a6ef62786fa7614"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.120326 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.122620 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" event={"ID":"6c720536-3f20-4c58-b914-ddbd3c52645b","Type":"ContainerStarted","Data":"71fe8c2722e0e385bc4b6288cb828577f374100cbc685e64428a694816d2a239"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.122661 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" event={"ID":"6c720536-3f20-4c58-b914-ddbd3c52645b","Type":"ContainerStarted","Data":"61d69e2408bafa4a9edb18a50b43117648429d3868f7ed9479b36fdb14af99c3"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.122902 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.157859 4837 generic.go:334] "Generic (PLEG): container finished" podID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerID="ce019f2b04dbaefe31b414be4203d9617d29bbb19556b4aef9311ab4c9d18e48" exitCode=1 Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.158047 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566cd58fb5-4lrgf" event={"ID":"75d8189f-c5f0-4328-b7ca-c979882a0b97","Type":"ContainerDied","Data":"ce019f2b04dbaefe31b414be4203d9617d29bbb19556b4aef9311ab4c9d18e48"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.158882 4837 scope.go:117] "RemoveContainer" containerID="ce019f2b04dbaefe31b414be4203d9617d29bbb19556b4aef9311ab4c9d18e48" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.159428 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerID="02f2d2c9a6bcb5aa3edee55129c8c41bcf6839c78b9645afd18010e96ab3db9c" exitCode=1 Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.160085 4837 scope.go:117] "RemoveContainer" containerID="02f2d2c9a6bcb5aa3edee55129c8c41bcf6839c78b9645afd18010e96ab3db9c" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.160302 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-844c4df74c-t57kg" event={"ID":"3f236154-f3c6-4347-aaa1-f0c338c4864d","Type":"ContainerDied","Data":"02f2d2c9a6bcb5aa3edee55129c8c41bcf6839c78b9645afd18010e96ab3db9c"} Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.160327 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.237737 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5dc5f6c468-wpmkd" podStartSLOduration=2.237714703 podStartE2EDuration="2.237714703s" podCreationTimestamp="2025-10-01 09:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:46.188717279 +0000 UTC m=+7623.030324734" watchObservedRunningTime="2025-10-01 09:12:46.237714703 +0000 UTC m=+7623.079322158" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.283271 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" podStartSLOduration=2.283256652 podStartE2EDuration="2.283256652s" podCreationTimestamp="2025-10-01 09:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:12:46.253562486 +0000 UTC m=+7623.095169941" watchObservedRunningTime="2025-10-01 09:12:46.283256652 +0000 UTC m=+7623.124864097" Oct 01 09:12:46 crc kubenswrapper[4837]: I1001 09:12:46.973751 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.113:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8443: connect: connection refused" Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.170665 4837 generic.go:334] "Generic (PLEG): container finished" podID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerID="9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a" exitCode=1 Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.170801 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566cd58fb5-4lrgf" event={"ID":"75d8189f-c5f0-4328-b7ca-c979882a0b97","Type":"ContainerDied","Data":"9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a"} Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.170844 4837 scope.go:117] "RemoveContainer" containerID="ce019f2b04dbaefe31b414be4203d9617d29bbb19556b4aef9311ab4c9d18e48" Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.171479 4837 scope.go:117] "RemoveContainer" containerID="9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a" Oct 01 09:12:47 crc kubenswrapper[4837]: E1001 09:12:47.171933 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566cd58fb5-4lrgf_openstack(75d8189f-c5f0-4328-b7ca-c979882a0b97)\"" pod="openstack/heat-api-566cd58fb5-4lrgf" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.175052 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerID="65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4" exitCode=1 Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.176513 4837 scope.go:117] "RemoveContainer" containerID="65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4" Oct 01 09:12:47 crc kubenswrapper[4837]: E1001 09:12:47.176848 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-844c4df74c-t57kg_openstack(3f236154-f3c6-4347-aaa1-f0c338c4864d)\"" pod="openstack/heat-cfnapi-844c4df74c-t57kg" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.176909 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-844c4df74c-t57kg" event={"ID":"3f236154-f3c6-4347-aaa1-f0c338c4864d","Type":"ContainerDied","Data":"65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4"} Oct 01 09:12:47 crc kubenswrapper[4837]: I1001 09:12:47.260963 4837 scope.go:117] "RemoveContainer" containerID="02f2d2c9a6bcb5aa3edee55129c8c41bcf6839c78b9645afd18010e96ab3db9c" Oct 01 09:12:48 crc kubenswrapper[4837]: I1001 09:12:48.188028 4837 scope.go:117] "RemoveContainer" containerID="9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a" Oct 01 09:12:48 crc kubenswrapper[4837]: E1001 09:12:48.188975 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566cd58fb5-4lrgf_openstack(75d8189f-c5f0-4328-b7ca-c979882a0b97)\"" pod="openstack/heat-api-566cd58fb5-4lrgf" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" Oct 01 09:12:48 crc kubenswrapper[4837]: I1001 09:12:48.190373 4837 scope.go:117] "RemoveContainer" containerID="65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4" Oct 01 09:12:48 crc kubenswrapper[4837]: E1001 09:12:48.190848 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-844c4df74c-t57kg_openstack(3f236154-f3c6-4347-aaa1-f0c338c4864d)\"" pod="openstack/heat-cfnapi-844c4df74c-t57kg" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.035722 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.036166 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.061561 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.061629 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.205067 4837 scope.go:117] "RemoveContainer" containerID="9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a" Oct 01 09:12:49 crc kubenswrapper[4837]: I1001 09:12:49.205232 4837 scope.go:117] "RemoveContainer" containerID="65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4" Oct 01 09:12:49 crc kubenswrapper[4837]: E1001 09:12:49.205620 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-844c4df74c-t57kg_openstack(3f236154-f3c6-4347-aaa1-f0c338c4864d)\"" pod="openstack/heat-cfnapi-844c4df74c-t57kg" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" Oct 01 09:12:49 crc kubenswrapper[4837]: E1001 09:12:49.205656 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566cd58fb5-4lrgf_openstack(75d8189f-c5f0-4328-b7ca-c979882a0b97)\"" pod="openstack/heat-api-566cd58fb5-4lrgf" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.168451 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-7f88999dcd-56w72" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.122:8004/healthcheck\": read tcp 10.217.0.2:50594->10.217.1.122:8004: read: connection reset by peer" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.227462 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.121:8000/healthcheck\": read tcp 10.217.0.2:39412->10.217.1.121:8000: read: connection reset by peer" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.678814 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.738026 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774572 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data\") pod \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774643 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle\") pod \"3446a7a3-bb73-48be-ad29-734880b2426e\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774711 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom\") pod \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774775 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb54x\" (UniqueName: \"kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x\") pod \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774811 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle\") pod \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\" (UID: \"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.774974 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrng2\" (UniqueName: \"kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2\") pod \"3446a7a3-bb73-48be-ad29-734880b2426e\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.775078 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom\") pod \"3446a7a3-bb73-48be-ad29-734880b2426e\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.775117 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data\") pod \"3446a7a3-bb73-48be-ad29-734880b2426e\" (UID: \"3446a7a3-bb73-48be-ad29-734880b2426e\") " Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.779982 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2" (OuterVolumeSpecName: "kube-api-access-qrng2") pod "3446a7a3-bb73-48be-ad29-734880b2426e" (UID: "3446a7a3-bb73-48be-ad29-734880b2426e"). InnerVolumeSpecName "kube-api-access-qrng2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.784882 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" (UID: "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.785822 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x" (OuterVolumeSpecName: "kube-api-access-mb54x") pod "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" (UID: "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b"). InnerVolumeSpecName "kube-api-access-mb54x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.788072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3446a7a3-bb73-48be-ad29-734880b2426e" (UID: "3446a7a3-bb73-48be-ad29-734880b2426e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.805862 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3446a7a3-bb73-48be-ad29-734880b2426e" (UID: "3446a7a3-bb73-48be-ad29-734880b2426e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.819492 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" (UID: "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.835347 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data" (OuterVolumeSpecName: "config-data") pod "3446a7a3-bb73-48be-ad29-734880b2426e" (UID: "3446a7a3-bb73-48be-ad29-734880b2426e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.840104 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data" (OuterVolumeSpecName: "config-data") pod "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" (UID: "3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877868 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrng2\" (UniqueName: \"kubernetes.io/projected/3446a7a3-bb73-48be-ad29-734880b2426e-kube-api-access-qrng2\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877894 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877904 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877913 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877921 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3446a7a3-bb73-48be-ad29-734880b2426e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877929 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877939 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb54x\" (UniqueName: \"kubernetes.io/projected/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-kube-api-access-mb54x\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:50 crc kubenswrapper[4837]: I1001 09:12:50.877948 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.223942 4837 generic.go:334] "Generic (PLEG): container finished" podID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerID="d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a" exitCode=0 Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.223983 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f88999dcd-56w72" event={"ID":"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b","Type":"ContainerDied","Data":"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a"} Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.224016 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7f88999dcd-56w72" event={"ID":"3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b","Type":"ContainerDied","Data":"ddc15214c960b8dbb4c8f0764c1c3ab20303404744ba1fdc55ccfbe53cda4153"} Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.224033 4837 scope.go:117] "RemoveContainer" containerID="d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.224090 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7f88999dcd-56w72" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.226058 4837 generic.go:334] "Generic (PLEG): container finished" podID="3446a7a3-bb73-48be-ad29-734880b2426e" containerID="01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a" exitCode=0 Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.226077 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" event={"ID":"3446a7a3-bb73-48be-ad29-734880b2426e","Type":"ContainerDied","Data":"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a"} Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.226094 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" event={"ID":"3446a7a3-bb73-48be-ad29-734880b2426e","Type":"ContainerDied","Data":"63a6f5d253b093993b0a1e1e7997f767208022fdce9623fb2979d4975ce9a5cd"} Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.226178 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c5bdbcc6-znf2l" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.257650 4837 scope.go:117] "RemoveContainer" containerID="d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a" Oct 01 09:12:51 crc kubenswrapper[4837]: E1001 09:12:51.258300 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a\": container with ID starting with d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a not found: ID does not exist" containerID="d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.258339 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a"} err="failed to get container status \"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a\": rpc error: code = NotFound desc = could not find container \"d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a\": container with ID starting with d2b04205bede3d118863caceb6cc73bbcca299c0722d4ae82a713e0d2990ed8a not found: ID does not exist" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.258362 4837 scope.go:117] "RemoveContainer" containerID="01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.281404 4837 scope.go:117] "RemoveContainer" containerID="01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a" Oct 01 09:12:51 crc kubenswrapper[4837]: E1001 09:12:51.281765 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a\": container with ID starting with 01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a not found: ID does not exist" containerID="01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.281813 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a"} err="failed to get container status \"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a\": rpc error: code = NotFound desc = could not find container \"01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a\": container with ID starting with 01fa685871cf23ef850865a49f5b2cc53afd0d3960f7f1698cf18547094b9c7a not found: ID does not exist" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.285734 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.297785 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7f88999dcd-56w72"] Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.306627 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.314175 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-8c5bdbcc6-znf2l"] Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.520451 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5dc5f6c468-wpmkd" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.574442 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.848674 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" path="/var/lib/kubelet/pods/3446a7a3-bb73-48be-ad29-734880b2426e/volumes" Oct 01 09:12:51 crc kubenswrapper[4837]: I1001 09:12:51.850417 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" path="/var/lib/kubelet/pods/3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b/volumes" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.020314 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.099097 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom\") pod \"75d8189f-c5f0-4328-b7ca-c979882a0b97\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.099289 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lz2j\" (UniqueName: \"kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j\") pod \"75d8189f-c5f0-4328-b7ca-c979882a0b97\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.099329 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data\") pod \"75d8189f-c5f0-4328-b7ca-c979882a0b97\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.099356 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle\") pod \"75d8189f-c5f0-4328-b7ca-c979882a0b97\" (UID: \"75d8189f-c5f0-4328-b7ca-c979882a0b97\") " Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.112249 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75d8189f-c5f0-4328-b7ca-c979882a0b97" (UID: "75d8189f-c5f0-4328-b7ca-c979882a0b97"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.136929 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j" (OuterVolumeSpecName: "kube-api-access-8lz2j") pod "75d8189f-c5f0-4328-b7ca-c979882a0b97" (UID: "75d8189f-c5f0-4328-b7ca-c979882a0b97"). InnerVolumeSpecName "kube-api-access-8lz2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.189313 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75d8189f-c5f0-4328-b7ca-c979882a0b97" (UID: "75d8189f-c5f0-4328-b7ca-c979882a0b97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.201423 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lz2j\" (UniqueName: \"kubernetes.io/projected/75d8189f-c5f0-4328-b7ca-c979882a0b97-kube-api-access-8lz2j\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.201458 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.201469 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.221630 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data" (OuterVolumeSpecName: "config-data") pod "75d8189f-c5f0-4328-b7ca-c979882a0b97" (UID: "75d8189f-c5f0-4328-b7ca-c979882a0b97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.236102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566cd58fb5-4lrgf" event={"ID":"75d8189f-c5f0-4328-b7ca-c979882a0b97","Type":"ContainerDied","Data":"08ad549f9800a176cfdb7407fb2ba9aa6047b1ccfdedfabb69149c12b6976496"} Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.236148 4837 scope.go:117] "RemoveContainer" containerID="9e080429f425381d4396b57a5c50f09b01e1586cd43381652f68767efbd3cc4a" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.236225 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566cd58fb5-4lrgf" Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.264390 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.282937 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-566cd58fb5-4lrgf"] Oct 01 09:12:52 crc kubenswrapper[4837]: I1001 09:12:52.302281 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d8189f-c5f0-4328-b7ca-c979882a0b97-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:53 crc kubenswrapper[4837]: I1001 09:12:53.828812 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" path="/var/lib/kubelet/pods/75d8189f-c5f0-4328-b7ca-c979882a0b97/volumes" Oct 01 09:12:56 crc kubenswrapper[4837]: I1001 09:12:56.695813 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-77ffb6c584-qdkpl" Oct 01 09:12:56 crc kubenswrapper[4837]: I1001 09:12:56.776562 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:56 crc kubenswrapper[4837]: I1001 09:12:56.974271 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bdcd979b8-z6dmt" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.113:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.113:8443: connect: connection refused" Oct 01 09:12:56 crc kubenswrapper[4837]: I1001 09:12:56.974412 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.178324 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.319499 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle\") pod \"3f236154-f3c6-4347-aaa1-f0c338c4864d\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.319572 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l5j8\" (UniqueName: \"kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8\") pod \"3f236154-f3c6-4347-aaa1-f0c338c4864d\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.319647 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data\") pod \"3f236154-f3c6-4347-aaa1-f0c338c4864d\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.319669 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom\") pod \"3f236154-f3c6-4347-aaa1-f0c338c4864d\" (UID: \"3f236154-f3c6-4347-aaa1-f0c338c4864d\") " Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.325294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-844c4df74c-t57kg" event={"ID":"3f236154-f3c6-4347-aaa1-f0c338c4864d","Type":"ContainerDied","Data":"7f6a38106a076d4e70173bc22ee07ca192376d51e752d7cba984832a762d5496"} Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.325350 4837 scope.go:117] "RemoveContainer" containerID="65d583e2473788e9a628a47c77ca6acef4a25bd79270128ad1af671cf115c7c4" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.325451 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-844c4df74c-t57kg" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.327122 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8" (OuterVolumeSpecName: "kube-api-access-2l5j8") pod "3f236154-f3c6-4347-aaa1-f0c338c4864d" (UID: "3f236154-f3c6-4347-aaa1-f0c338c4864d"). InnerVolumeSpecName "kube-api-access-2l5j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.331017 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3f236154-f3c6-4347-aaa1-f0c338c4864d" (UID: "3f236154-f3c6-4347-aaa1-f0c338c4864d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.355340 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f236154-f3c6-4347-aaa1-f0c338c4864d" (UID: "3f236154-f3c6-4347-aaa1-f0c338c4864d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.377026 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data" (OuterVolumeSpecName: "config-data") pod "3f236154-f3c6-4347-aaa1-f0c338c4864d" (UID: "3f236154-f3c6-4347-aaa1-f0c338c4864d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.422851 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.423056 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.423137 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f236154-f3c6-4347-aaa1-f0c338c4864d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.423244 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l5j8\" (UniqueName: \"kubernetes.io/projected/3f236154-f3c6-4347-aaa1-f0c338c4864d-kube-api-access-2l5j8\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.580436 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.677745 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.687457 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-844c4df74c-t57kg"] Oct 01 09:12:57 crc kubenswrapper[4837]: I1001 09:12:57.834500 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" path="/var/lib/kubelet/pods/3f236154-f3c6-4347-aaa1-f0c338c4864d/volumes" Oct 01 09:12:57 crc kubenswrapper[4837]: E1001 09:12:57.849790 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f236154_f3c6_4347_aaa1_f0c338c4864d.slice\": RecentStats: unable to find data in memory cache]" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.346305 4837 generic.go:334] "Generic (PLEG): container finished" podID="895440de-0fde-4148-9c31-4547b9c12764" containerID="99fdcae14a1692b6ffa7e89638b0b14c7f8c0504163cdf0d012c71fb5f50a9f2" exitCode=137 Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.346528 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerDied","Data":"99fdcae14a1692b6ffa7e89638b0b14c7f8c0504163cdf0d012c71fb5f50a9f2"} Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.347042 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdcd979b8-z6dmt" event={"ID":"895440de-0fde-4148-9c31-4547b9c12764","Type":"ContainerDied","Data":"25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316"} Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.347061 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25e968bcb73c9ccdbe0577a7230f81ae8c770a767206678000f27485d1916316" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.420217 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.570846 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.570896 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571019 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571121 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571149 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvvj4\" (UniqueName: \"kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571166 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571181 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key\") pod \"895440de-0fde-4148-9c31-4547b9c12764\" (UID: \"895440de-0fde-4148-9c31-4547b9c12764\") " Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.571753 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs" (OuterVolumeSpecName: "logs") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.577210 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.583069 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4" (OuterVolumeSpecName: "kube-api-access-lvvj4") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "kube-api-access-lvvj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.607311 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts" (OuterVolumeSpecName: "scripts") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.610970 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.621982 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data" (OuterVolumeSpecName: "config-data") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.661838 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "895440de-0fde-4148-9c31-4547b9c12764" (UID: "895440de-0fde-4148-9c31-4547b9c12764"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673042 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673078 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvvj4\" (UniqueName: \"kubernetes.io/projected/895440de-0fde-4148-9c31-4547b9c12764-kube-api-access-lvvj4\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673092 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673105 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673115 4837 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/895440de-0fde-4148-9c31-4547b9c12764-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673125 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/895440de-0fde-4148-9c31-4547b9c12764-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:12:59 crc kubenswrapper[4837]: I1001 09:12:59.673134 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/895440de-0fde-4148-9c31-4547b9c12764-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:00 crc kubenswrapper[4837]: I1001 09:13:00.356004 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdcd979b8-z6dmt" Oct 01 09:13:00 crc kubenswrapper[4837]: I1001 09:13:00.389632 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:13:00 crc kubenswrapper[4837]: I1001 09:13:00.401231 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bdcd979b8-z6dmt"] Oct 01 09:13:01 crc kubenswrapper[4837]: I1001 09:13:01.836918 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895440de-0fde-4148-9c31-4547b9c12764" path="/var/lib/kubelet/pods/895440de-0fde-4148-9c31-4547b9c12764/volumes" Oct 01 09:13:03 crc kubenswrapper[4837]: I1001 09:13:03.988155 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-b8cc69c8f-zc64n" Oct 01 09:13:04 crc kubenswrapper[4837]: I1001 09:13:04.038684 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:13:04 crc kubenswrapper[4837]: I1001 09:13:04.039148 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-678489d5c5-5rs4j" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerName="heat-engine" containerID="cri-o://8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" gracePeriod=60 Oct 01 09:13:07 crc kubenswrapper[4837]: E1001 09:13:07.527760 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 01 09:13:07 crc kubenswrapper[4837]: E1001 09:13:07.529856 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 01 09:13:07 crc kubenswrapper[4837]: E1001 09:13:07.531367 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 01 09:13:07 crc kubenswrapper[4837]: E1001 09:13:07.531413 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-678489d5c5-5rs4j" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerName="heat-engine" Oct 01 09:13:11 crc kubenswrapper[4837]: I1001 09:13:11.000478 4837 scope.go:117] "RemoveContainer" containerID="e653f71333e1d55c71b277b79feb0e810a2ff196e85a96d4baded6dae9104645" Oct 01 09:13:11 crc kubenswrapper[4837]: I1001 09:13:11.035415 4837 scope.go:117] "RemoveContainer" containerID="ebf9bb932aa70cb2e5313374790bf84e50d8593ca7e4969f1cc332ce43c99342" Oct 01 09:13:11 crc kubenswrapper[4837]: I1001 09:13:11.101421 4837 scope.go:117] "RemoveContainer" containerID="d41c236862bbb1755f8ba5864b9eca5ca1bd474ab6b95aebbf365a136c0b1559" Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.521798 4837 generic.go:334] "Generic (PLEG): container finished" podID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerID="8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" exitCode=0 Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.521885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678489d5c5-5rs4j" event={"ID":"c685b9c0-c2df-4408-8fd9-9bebdf59b732","Type":"ContainerDied","Data":"8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f"} Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.797390 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.969443 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb6gg\" (UniqueName: \"kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg\") pod \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.969570 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data\") pod \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.969648 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom\") pod \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.969886 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle\") pod \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\" (UID: \"c685b9c0-c2df-4408-8fd9-9bebdf59b732\") " Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.977557 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c685b9c0-c2df-4408-8fd9-9bebdf59b732" (UID: "c685b9c0-c2df-4408-8fd9-9bebdf59b732"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:16 crc kubenswrapper[4837]: I1001 09:13:16.978915 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg" (OuterVolumeSpecName: "kube-api-access-gb6gg") pod "c685b9c0-c2df-4408-8fd9-9bebdf59b732" (UID: "c685b9c0-c2df-4408-8fd9-9bebdf59b732"). InnerVolumeSpecName "kube-api-access-gb6gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.008793 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c685b9c0-c2df-4408-8fd9-9bebdf59b732" (UID: "c685b9c0-c2df-4408-8fd9-9bebdf59b732"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.045886 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data" (OuterVolumeSpecName: "config-data") pod "c685b9c0-c2df-4408-8fd9-9bebdf59b732" (UID: "c685b9c0-c2df-4408-8fd9-9bebdf59b732"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.072448 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.072523 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb6gg\" (UniqueName: \"kubernetes.io/projected/c685b9c0-c2df-4408-8fd9-9bebdf59b732-kube-api-access-gb6gg\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.072535 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.072546 4837 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c685b9c0-c2df-4408-8fd9-9bebdf59b732-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.539926 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678489d5c5-5rs4j" event={"ID":"c685b9c0-c2df-4408-8fd9-9bebdf59b732","Type":"ContainerDied","Data":"ff94b1880ffa3894afe855b6baaab988581015dace4fa9bcc3daca8b0f6881e7"} Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.539989 4837 scope.go:117] "RemoveContainer" containerID="8643588c68d25ec2a16e0b39be7ebbe0443e355f4b17e7a54770eff3e22ecb1f" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.540001 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678489d5c5-5rs4j" Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.596564 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.610140 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-678489d5c5-5rs4j"] Oct 01 09:13:17 crc kubenswrapper[4837]: I1001 09:13:17.834263 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" path="/var/lib/kubelet/pods/c685b9c0-c2df-4408-8fd9-9bebdf59b732/volumes" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.604765 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc"] Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605232 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605245 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605260 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerName="heat-engine" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605267 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerName="heat-engine" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605283 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605289 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605303 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605308 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605319 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605324 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605337 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605343 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605352 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605358 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.605365 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon-log" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605370 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon-log" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605532 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605545 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c685b9c0-c2df-4408-8fd9-9bebdf59b732" containerName="heat-engine" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605558 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f236154-f3c6-4347-aaa1-f0c338c4864d" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605572 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8acf25-f7b4-4af5-a0e6-ca11cc839a3b" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605581 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605592 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605601 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605609 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="895440de-0fde-4148-9c31-4547b9c12764" containerName="horizon-log" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.605616 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3446a7a3-bb73-48be-ad29-734880b2426e" containerName="heat-cfnapi" Oct 01 09:13:18 crc kubenswrapper[4837]: E1001 09:13:18.606316 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.606327 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d8189f-c5f0-4328-b7ca-c979882a0b97" containerName="heat-api" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.607475 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.610105 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.638815 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc"] Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.713233 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.713408 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.713457 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sksdx\" (UniqueName: \"kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.815943 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.816047 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sksdx\" (UniqueName: \"kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.816114 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.816628 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.816804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.847899 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sksdx\" (UniqueName: \"kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:18 crc kubenswrapper[4837]: I1001 09:13:18.945466 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:19 crc kubenswrapper[4837]: I1001 09:13:19.481329 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc"] Oct 01 09:13:19 crc kubenswrapper[4837]: I1001 09:13:19.569759 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" event={"ID":"a9cdebd9-f254-406b-9e2a-775993bc8b11","Type":"ContainerStarted","Data":"3869d470a4554ff0a3030eac0f638ea1f70c4a39c81aeab1b2f90dbc878aa9e8"} Oct 01 09:13:20 crc kubenswrapper[4837]: I1001 09:13:20.584950 4837 generic.go:334] "Generic (PLEG): container finished" podID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerID="c2b4c81d9cd9fecdc73e53610368c475686e770520853c80ded7ff8eba5b7bb6" exitCode=0 Oct 01 09:13:20 crc kubenswrapper[4837]: I1001 09:13:20.585016 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" event={"ID":"a9cdebd9-f254-406b-9e2a-775993bc8b11","Type":"ContainerDied","Data":"c2b4c81d9cd9fecdc73e53610368c475686e770520853c80ded7ff8eba5b7bb6"} Oct 01 09:13:22 crc kubenswrapper[4837]: I1001 09:13:22.618808 4837 generic.go:334] "Generic (PLEG): container finished" podID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerID="c99df9ceb06c56f001cde5254003b728c40c43d81cb06b744f6f92ed73779d35" exitCode=0 Oct 01 09:13:22 crc kubenswrapper[4837]: I1001 09:13:22.619002 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" event={"ID":"a9cdebd9-f254-406b-9e2a-775993bc8b11","Type":"ContainerDied","Data":"c99df9ceb06c56f001cde5254003b728c40c43d81cb06b744f6f92ed73779d35"} Oct 01 09:13:23 crc kubenswrapper[4837]: I1001 09:13:23.079069 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:13:23 crc kubenswrapper[4837]: I1001 09:13:23.079344 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:13:23 crc kubenswrapper[4837]: I1001 09:13:23.645203 4837 generic.go:334] "Generic (PLEG): container finished" podID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerID="ca8a22829aadc5f76a2a01eabb1fe76a6823582484c7120b22c36cc2ddae6967" exitCode=0 Oct 01 09:13:23 crc kubenswrapper[4837]: I1001 09:13:23.645287 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" event={"ID":"a9cdebd9-f254-406b-9e2a-775993bc8b11","Type":"ContainerDied","Data":"ca8a22829aadc5f76a2a01eabb1fe76a6823582484c7120b22c36cc2ddae6967"} Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.120427 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.261769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sksdx\" (UniqueName: \"kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx\") pod \"a9cdebd9-f254-406b-9e2a-775993bc8b11\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.261819 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle\") pod \"a9cdebd9-f254-406b-9e2a-775993bc8b11\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.261864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util\") pod \"a9cdebd9-f254-406b-9e2a-775993bc8b11\" (UID: \"a9cdebd9-f254-406b-9e2a-775993bc8b11\") " Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.263823 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle" (OuterVolumeSpecName: "bundle") pod "a9cdebd9-f254-406b-9e2a-775993bc8b11" (UID: "a9cdebd9-f254-406b-9e2a-775993bc8b11"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.267528 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx" (OuterVolumeSpecName: "kube-api-access-sksdx") pod "a9cdebd9-f254-406b-9e2a-775993bc8b11" (UID: "a9cdebd9-f254-406b-9e2a-775993bc8b11"). InnerVolumeSpecName "kube-api-access-sksdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.273458 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util" (OuterVolumeSpecName: "util") pod "a9cdebd9-f254-406b-9e2a-775993bc8b11" (UID: "a9cdebd9-f254-406b-9e2a-775993bc8b11"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.364931 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sksdx\" (UniqueName: \"kubernetes.io/projected/a9cdebd9-f254-406b-9e2a-775993bc8b11-kube-api-access-sksdx\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.364970 4837 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.364982 4837 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9cdebd9-f254-406b-9e2a-775993bc8b11-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.695547 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" event={"ID":"a9cdebd9-f254-406b-9e2a-775993bc8b11","Type":"ContainerDied","Data":"3869d470a4554ff0a3030eac0f638ea1f70c4a39c81aeab1b2f90dbc878aa9e8"} Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.695615 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3869d470a4554ff0a3030eac0f638ea1f70c4a39c81aeab1b2f90dbc878aa9e8" Oct 01 09:13:25 crc kubenswrapper[4837]: I1001 09:13:25.695771 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.880190 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz"] Oct 01 09:13:35 crc kubenswrapper[4837]: E1001 09:13:35.880955 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="util" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.881126 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="util" Oct 01 09:13:35 crc kubenswrapper[4837]: E1001 09:13:35.881143 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="extract" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.881150 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="extract" Oct 01 09:13:35 crc kubenswrapper[4837]: E1001 09:13:35.881168 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="pull" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.881174 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="pull" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.881341 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cdebd9-f254-406b-9e2a-775993bc8b11" containerName="extract" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.881976 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.884204 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.884401 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-jpd5w" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.884782 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 01 09:13:35 crc kubenswrapper[4837]: I1001 09:13:35.935560 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.018854 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g2vc\" (UniqueName: \"kubernetes.io/projected/90905d8d-f989-49a4-9f32-2648922f050b-kube-api-access-2g2vc\") pod \"obo-prometheus-operator-7c8cf85677-vh2dz\" (UID: \"90905d8d-f989-49a4-9f32-2648922f050b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.066815 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.068008 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.070400 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-8jsrs" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.073459 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.081428 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.082708 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.095885 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.120160 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g2vc\" (UniqueName: \"kubernetes.io/projected/90905d8d-f989-49a4-9f32-2648922f050b-kube-api-access-2g2vc\") pod \"obo-prometheus-operator-7c8cf85677-vh2dz\" (UID: \"90905d8d-f989-49a4-9f32-2648922f050b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.128813 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.167455 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g2vc\" (UniqueName: \"kubernetes.io/projected/90905d8d-f989-49a4-9f32-2648922f050b-kube-api-access-2g2vc\") pod \"obo-prometheus-operator-7c8cf85677-vh2dz\" (UID: \"90905d8d-f989-49a4-9f32-2648922f050b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.202323 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.224013 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.224079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.224187 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.224230 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.273654 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-q7w4h"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.274909 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.278053 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.281788 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cntpc" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.298167 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-q7w4h"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.328936 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.328998 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.329073 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.329106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.333602 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.333873 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10948123-507c-4b8b-9333-0a29d502537e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk\" (UID: \"10948123-507c-4b8b-9333-0a29d502537e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.334357 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.335034 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a2fc8670-131c-486d-800e-57b543426d8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz\" (UID: \"a2fc8670-131c-486d-800e-57b543426d8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.383148 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.385200 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-jmfpx"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.386367 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.393224 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-xgf4s" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.399249 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.419140 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-jmfpx"] Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.431189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7245\" (UniqueName: \"kubernetes.io/projected/2881b32b-58f6-4790-9011-82e66406b813-kube-api-access-n7245\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.431240 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2881b32b-58f6-4790-9011-82e66406b813-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.536314 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.536649 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pr45\" (UniqueName: \"kubernetes.io/projected/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-kube-api-access-5pr45\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.536682 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7245\" (UniqueName: \"kubernetes.io/projected/2881b32b-58f6-4790-9011-82e66406b813-kube-api-access-n7245\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.536721 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2881b32b-58f6-4790-9011-82e66406b813-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.546500 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/2881b32b-58f6-4790-9011-82e66406b813-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.561586 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7245\" (UniqueName: \"kubernetes.io/projected/2881b32b-58f6-4790-9011-82e66406b813-kube-api-access-n7245\") pod \"observability-operator-cc5f78dfc-q7w4h\" (UID: \"2881b32b-58f6-4790-9011-82e66406b813\") " pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.638089 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.638257 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pr45\" (UniqueName: \"kubernetes.io/projected/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-kube-api-access-5pr45\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.639110 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.656097 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pr45\" (UniqueName: \"kubernetes.io/projected/ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb-kube-api-access-5pr45\") pod \"perses-operator-54bc95c9fb-jmfpx\" (UID: \"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb\") " pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.747751 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.764355 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:36 crc kubenswrapper[4837]: I1001 09:13:36.820819 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz"] Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.009786 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk"] Oct 01 09:13:37 crc kubenswrapper[4837]: W1001 09:13:37.015062 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10948123_507c_4b8b_9333_0a29d502537e.slice/crio-5c8471ac3ecf0aed59945b2664cf0b467e6261341c3d2e85d4ed5e63c913db36 WatchSource:0}: Error finding container 5c8471ac3ecf0aed59945b2664cf0b467e6261341c3d2e85d4ed5e63c913db36: Status 404 returned error can't find the container with id 5c8471ac3ecf0aed59945b2664cf0b467e6261341c3d2e85d4ed5e63c913db36 Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.110566 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz"] Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.287208 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-q7w4h"] Oct 01 09:13:37 crc kubenswrapper[4837]: W1001 09:13:37.291849 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2881b32b_58f6_4790_9011_82e66406b813.slice/crio-8e7b6427656cc9527074e311d219a752f810fbf79547b75166692b07b1c0bebd WatchSource:0}: Error finding container 8e7b6427656cc9527074e311d219a752f810fbf79547b75166692b07b1c0bebd: Status 404 returned error can't find the container with id 8e7b6427656cc9527074e311d219a752f810fbf79547b75166692b07b1c0bebd Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.372726 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-jmfpx"] Oct 01 09:13:37 crc kubenswrapper[4837]: W1001 09:13:37.375795 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff0addb2_dc6a_4838_ac75_e5cbcd7d33eb.slice/crio-6593fcd038eb90eda33a21ac974204a1737af3c90ce7de5b50ca549ff1c705d1 WatchSource:0}: Error finding container 6593fcd038eb90eda33a21ac974204a1737af3c90ce7de5b50ca549ff1c705d1: Status 404 returned error can't find the container with id 6593fcd038eb90eda33a21ac974204a1737af3c90ce7de5b50ca549ff1c705d1 Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.911004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" event={"ID":"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb","Type":"ContainerStarted","Data":"6593fcd038eb90eda33a21ac974204a1737af3c90ce7de5b50ca549ff1c705d1"} Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.922936 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" event={"ID":"2881b32b-58f6-4790-9011-82e66406b813","Type":"ContainerStarted","Data":"8e7b6427656cc9527074e311d219a752f810fbf79547b75166692b07b1c0bebd"} Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.936058 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" event={"ID":"90905d8d-f989-49a4-9f32-2648922f050b","Type":"ContainerStarted","Data":"2a78acc11b321600b09b453b0b9121290bdff929808376d1dbe6e348fc13f55f"} Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.950055 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" event={"ID":"a2fc8670-131c-486d-800e-57b543426d8c","Type":"ContainerStarted","Data":"0f1c427bf60211b34287f4f266d3f1680479408e6ae497864350fccc92551960"} Oct 01 09:13:37 crc kubenswrapper[4837]: I1001 09:13:37.959835 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" event={"ID":"10948123-507c-4b8b-9333-0a29d502537e","Type":"ContainerStarted","Data":"5c8471ac3ecf0aed59945b2664cf0b467e6261341c3d2e85d4ed5e63c913db36"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.089623 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" event={"ID":"2881b32b-58f6-4790-9011-82e66406b813","Type":"ContainerStarted","Data":"a918f37028e82aa366cada5dcd96c9c7a179be389db8e6e3ff417b1b73c6b0b0"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.090279 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.091265 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" event={"ID":"90905d8d-f989-49a4-9f32-2648922f050b","Type":"ContainerStarted","Data":"4877f2e46d3f21565557eae07df7e7eba99787f4fe61beb3b586b778b8f3067f"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.092985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" event={"ID":"a2fc8670-131c-486d-800e-57b543426d8c","Type":"ContainerStarted","Data":"e942071e98df6fed12893292bbb816ffba2914da16245da98d70ea0345881d8f"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.094308 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" event={"ID":"10948123-507c-4b8b-9333-0a29d502537e","Type":"ContainerStarted","Data":"a866b1a1eabf169632da205d41ac71c7755dbaebe27d4f374be61c137abc6513"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.095514 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" event={"ID":"ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb","Type":"ContainerStarted","Data":"786c3c83ff50c8b843db8722fd27577190870cb42fd5e98a0421c45ef453d5f5"} Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.096069 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.117769 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" podStartSLOduration=2.109876906 podStartE2EDuration="10.117748463s" podCreationTimestamp="2025-10-01 09:13:36 +0000 UTC" firstStartedPulling="2025-10-01 09:13:37.294059534 +0000 UTC m=+7674.135666989" lastFinishedPulling="2025-10-01 09:13:45.301931091 +0000 UTC m=+7682.143538546" observedRunningTime="2025-10-01 09:13:46.11516121 +0000 UTC m=+7682.956768665" watchObservedRunningTime="2025-10-01 09:13:46.117748463 +0000 UTC m=+7682.959355928" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.132749 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-vh2dz" podStartSLOduration=2.822915688 podStartE2EDuration="11.132732675s" podCreationTimestamp="2025-10-01 09:13:35 +0000 UTC" firstStartedPulling="2025-10-01 09:13:36.853854358 +0000 UTC m=+7673.695461813" lastFinishedPulling="2025-10-01 09:13:45.163671355 +0000 UTC m=+7682.005278800" observedRunningTime="2025-10-01 09:13:46.127346361 +0000 UTC m=+7682.968953816" watchObservedRunningTime="2025-10-01 09:13:46.132732675 +0000 UTC m=+7682.974340130" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.147654 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-q7w4h" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.185308 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" podStartSLOduration=2.402447436 podStartE2EDuration="10.185290198s" podCreationTimestamp="2025-10-01 09:13:36 +0000 UTC" firstStartedPulling="2025-10-01 09:13:37.378936097 +0000 UTC m=+7674.220543562" lastFinishedPulling="2025-10-01 09:13:45.161778879 +0000 UTC m=+7682.003386324" observedRunningTime="2025-10-01 09:13:46.179251297 +0000 UTC m=+7683.020858772" watchObservedRunningTime="2025-10-01 09:13:46.185290198 +0000 UTC m=+7683.026897663" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.191491 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk" podStartSLOduration=1.9954647429999999 podStartE2EDuration="10.1914724s" podCreationTimestamp="2025-10-01 09:13:36 +0000 UTC" firstStartedPulling="2025-10-01 09:13:37.019830071 +0000 UTC m=+7673.861437526" lastFinishedPulling="2025-10-01 09:13:45.215837718 +0000 UTC m=+7682.057445183" observedRunningTime="2025-10-01 09:13:46.15430472 +0000 UTC m=+7682.995912175" watchObservedRunningTime="2025-10-01 09:13:46.1914724 +0000 UTC m=+7683.033079875" Oct 01 09:13:46 crc kubenswrapper[4837]: I1001 09:13:46.200285 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz" podStartSLOduration=2.171286338 podStartE2EDuration="10.200268108s" podCreationTimestamp="2025-10-01 09:13:36 +0000 UTC" firstStartedPulling="2025-10-01 09:13:37.131832965 +0000 UTC m=+7673.973440420" lastFinishedPulling="2025-10-01 09:13:45.160814745 +0000 UTC m=+7682.002422190" observedRunningTime="2025-10-01 09:13:46.196633489 +0000 UTC m=+7683.038240944" watchObservedRunningTime="2025-10-01 09:13:46.200268108 +0000 UTC m=+7683.041875563" Oct 01 09:13:53 crc kubenswrapper[4837]: I1001 09:13:53.039052 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-96d22"] Oct 01 09:13:53 crc kubenswrapper[4837]: I1001 09:13:53.052819 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-96d22"] Oct 01 09:13:53 crc kubenswrapper[4837]: I1001 09:13:53.079855 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:13:53 crc kubenswrapper[4837]: I1001 09:13:53.080235 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:13:53 crc kubenswrapper[4837]: I1001 09:13:53.836609 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="007a2a06-3b17-43b4-b070-df7291499232" path="/var/lib/kubelet/pods/007a2a06-3b17-43b4-b070-df7291499232/volumes" Oct 01 09:13:56 crc kubenswrapper[4837]: I1001 09:13:56.820796 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-jmfpx" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.479778 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.490858 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="b327f5dc-78e7-412c-a281-d281511fefa5" containerName="openstackclient" containerID="cri-o://b7a9137cf9d14a8e88f10c96a379d917d6fc6c0b280bd6a9405e8d1a81622fe2" gracePeriod=2 Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.495322 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.554538 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 09:13:59 crc kubenswrapper[4837]: E1001 09:13:59.554970 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b327f5dc-78e7-412c-a281-d281511fefa5" containerName="openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.554987 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b327f5dc-78e7-412c-a281-d281511fefa5" containerName="openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.555156 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b327f5dc-78e7-412c-a281-d281511fefa5" containerName="openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.555793 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.578546 4837 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b327f5dc-78e7-412c-a281-d281511fefa5" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.602870 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.680916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.681293 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.681414 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.681530 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lp9n\" (UniqueName: \"kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.783687 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.783788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lp9n\" (UniqueName: \"kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.783820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.783843 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.787722 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.789440 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.795840 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.831806 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.833379 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.836481 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qbg56" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.874129 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lp9n\" (UniqueName: \"kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n\") pod \"openstackclient\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.978948 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.981521 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wct\" (UniqueName: \"kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct\") pod \"kube-state-metrics-0\" (UID: \"30220513-03a1-44c1-b372-b878b9701a6e\") " pod="openstack/kube-state-metrics-0" Oct 01 09:13:59 crc kubenswrapper[4837]: I1001 09:13:59.990036 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.083720 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wct\" (UniqueName: \"kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct\") pod \"kube-state-metrics-0\" (UID: \"30220513-03a1-44c1-b372-b878b9701a6e\") " pod="openstack/kube-state-metrics-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.101363 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wct\" (UniqueName: \"kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct\") pod \"kube-state-metrics-0\" (UID: \"30220513-03a1-44c1-b372-b878b9701a6e\") " pod="openstack/kube-state-metrics-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.327655 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.330616 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.334462 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.338497 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.338731 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.339531 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-78h2j" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.339657 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.343038 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.396849 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.396908 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w97xm\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-kube-api-access-w97xm\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.397021 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.397055 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.397103 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.397124 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498235 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498606 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498637 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498681 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w97xm\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-kube-api-access-w97xm\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.498950 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.507214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.507421 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.509200 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.513109 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3f958bc4-38ae-49d4-9a51-205ce949a3b8-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.533423 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w97xm\" (UniqueName: \"kubernetes.io/projected/3f958bc4-38ae-49d4-9a51-205ce949a3b8-kube-api-access-w97xm\") pod \"alertmanager-metric-storage-0\" (UID: \"3f958bc4-38ae-49d4-9a51-205ce949a3b8\") " pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.681261 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:00 crc kubenswrapper[4837]: I1001 09:14:00.861158 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.102800 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.105625 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.115297 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.122444 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.122625 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-g867p" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.123144 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.123274 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.139257 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.151329 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.156990 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.238398 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.239939 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240024 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240130 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240249 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240331 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpk2w\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.240523 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.290442 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30220513-03a1-44c1-b372-b878b9701a6e","Type":"ContainerStarted","Data":"5d17d62e7bcc3a23de4b5f8cafb630269c6671230701c2ea8b9c1cd842a9559d"} Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.291816 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dd1784a7-4be8-4517-a06e-2b3209cd67b8","Type":"ContainerStarted","Data":"ac61d35d9557a9ab47274dd971284105785d08d9b32d00dd3da8098fd66db8da"} Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.291858 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dd1784a7-4be8-4517-a06e-2b3209cd67b8","Type":"ContainerStarted","Data":"5b74f7bfafbc737ed5fe6c83d591334452089375d5bab04d502b17b1f6ce86c7"} Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.328166 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.328150364 podStartE2EDuration="2.328150364s" podCreationTimestamp="2025-10-01 09:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:14:01.327961989 +0000 UTC m=+7698.169569444" watchObservedRunningTime="2025-10-01 09:14:01.328150364 +0000 UTC m=+7698.169757819" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.342104 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.342576 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.342664 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.342787 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpk2w\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.342868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.343008 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.343119 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.343194 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.346402 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.349411 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.349541 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5a32693e0b704d1d24b4f326abcdde6b5965d6dd0bf58ad7942285c2e3caa226/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.351798 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.352450 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.353883 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.355430 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.365041 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.371795 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpk2w\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.499138 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.602936 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:01 crc kubenswrapper[4837]: I1001 09:14:01.722586 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.330637 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.354011 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30220513-03a1-44c1-b372-b878b9701a6e","Type":"ContainerStarted","Data":"fbc28bf12bdab8f685750f34986cdd0c7fd27df5a85c92937f42f44fbc0a414c"} Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.354298 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.373960 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"3f958bc4-38ae-49d4-9a51-205ce949a3b8","Type":"ContainerStarted","Data":"adb3892a0a7db3197ad5488e2b0a286675469988eef09bcefb0d60ae6905d172"} Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.398831 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.002275626 podStartE2EDuration="3.398808451s" podCreationTimestamp="2025-10-01 09:13:59 +0000 UTC" firstStartedPulling="2025-10-01 09:14:01.145946559 +0000 UTC m=+7697.987554014" lastFinishedPulling="2025-10-01 09:14:01.542479394 +0000 UTC m=+7698.384086839" observedRunningTime="2025-10-01 09:14:02.387211913 +0000 UTC m=+7699.228819368" watchObservedRunningTime="2025-10-01 09:14:02.398808451 +0000 UTC m=+7699.240415906" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.414272 4837 generic.go:334] "Generic (PLEG): container finished" podID="b327f5dc-78e7-412c-a281-d281511fefa5" containerID="b7a9137cf9d14a8e88f10c96a379d917d6fc6c0b280bd6a9405e8d1a81622fe2" exitCode=137 Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.414451 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cce82aa885899e50ddfdadf0932f615e7ba2cdfb9dab1fe874209b7389be821" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.427389 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.579427 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config\") pod \"b327f5dc-78e7-412c-a281-d281511fefa5\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.579607 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn7j7\" (UniqueName: \"kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7\") pod \"b327f5dc-78e7-412c-a281-d281511fefa5\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.579805 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle\") pod \"b327f5dc-78e7-412c-a281-d281511fefa5\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.579850 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret\") pod \"b327f5dc-78e7-412c-a281-d281511fefa5\" (UID: \"b327f5dc-78e7-412c-a281-d281511fefa5\") " Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.584849 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7" (OuterVolumeSpecName: "kube-api-access-jn7j7") pod "b327f5dc-78e7-412c-a281-d281511fefa5" (UID: "b327f5dc-78e7-412c-a281-d281511fefa5"). InnerVolumeSpecName "kube-api-access-jn7j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.609545 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b327f5dc-78e7-412c-a281-d281511fefa5" (UID: "b327f5dc-78e7-412c-a281-d281511fefa5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.614946 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b327f5dc-78e7-412c-a281-d281511fefa5" (UID: "b327f5dc-78e7-412c-a281-d281511fefa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.654905 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b327f5dc-78e7-412c-a281-d281511fefa5" (UID: "b327f5dc-78e7-412c-a281-d281511fefa5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.683188 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.683754 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b327f5dc-78e7-412c-a281-d281511fefa5-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.683765 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn7j7\" (UniqueName: \"kubernetes.io/projected/b327f5dc-78e7-412c-a281-d281511fefa5-kube-api-access-jn7j7\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:02 crc kubenswrapper[4837]: I1001 09:14:02.683773 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b327f5dc-78e7-412c-a281-d281511fefa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.045485 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-31a1-account-create-sfpml"] Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.054876 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-31a1-account-create-sfpml"] Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.423886 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.423918 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerStarted","Data":"6b4feaaf42665f3634469d84bd1c4d1859be92e640045b744ab91e9879abb832"} Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.440858 4837 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b327f5dc-78e7-412c-a281-d281511fefa5" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.830370 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="011cca03-8337-4be1-9b4d-c2d3862a0db5" path="/var/lib/kubelet/pods/011cca03-8337-4be1-9b4d-c2d3862a0db5/volumes" Oct 01 09:14:03 crc kubenswrapper[4837]: I1001 09:14:03.831647 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b327f5dc-78e7-412c-a281-d281511fefa5" path="/var/lib/kubelet/pods/b327f5dc-78e7-412c-a281-d281511fefa5/volumes" Oct 01 09:14:07 crc kubenswrapper[4837]: I1001 09:14:07.483001 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerStarted","Data":"349d0f74f3b69457bbf8cc775c808dcc432384da3de7ffe998cf510a4318c278"} Oct 01 09:14:08 crc kubenswrapper[4837]: I1001 09:14:08.499476 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"3f958bc4-38ae-49d4-9a51-205ce949a3b8","Type":"ContainerStarted","Data":"a4c5432f9fc74db9bdff070975158e6a095ff2ba0b3b32b32b90127446998f81"} Oct 01 09:14:10 crc kubenswrapper[4837]: I1001 09:14:10.340194 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 09:14:11 crc kubenswrapper[4837]: I1001 09:14:11.341008 4837 scope.go:117] "RemoveContainer" containerID="1d49f5cc9fdaf44790f2e29467bfe6ef969b66744323a2e05955d0ea02578963" Oct 01 09:14:11 crc kubenswrapper[4837]: I1001 09:14:11.376967 4837 scope.go:117] "RemoveContainer" containerID="b44dd1c021485d8bdfb1f9f176ab5210c897de7b373ac5af4b8548cc51edf723" Oct 01 09:14:11 crc kubenswrapper[4837]: I1001 09:14:11.445432 4837 scope.go:117] "RemoveContainer" containerID="b7a9137cf9d14a8e88f10c96a379d917d6fc6c0b280bd6a9405e8d1a81622fe2" Oct 01 09:14:16 crc kubenswrapper[4837]: I1001 09:14:16.587398 4837 generic.go:334] "Generic (PLEG): container finished" podID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerID="349d0f74f3b69457bbf8cc775c808dcc432384da3de7ffe998cf510a4318c278" exitCode=0 Oct 01 09:14:16 crc kubenswrapper[4837]: I1001 09:14:16.587542 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerDied","Data":"349d0f74f3b69457bbf8cc775c808dcc432384da3de7ffe998cf510a4318c278"} Oct 01 09:14:16 crc kubenswrapper[4837]: I1001 09:14:16.597026 4837 generic.go:334] "Generic (PLEG): container finished" podID="3f958bc4-38ae-49d4-9a51-205ce949a3b8" containerID="a4c5432f9fc74db9bdff070975158e6a095ff2ba0b3b32b32b90127446998f81" exitCode=0 Oct 01 09:14:16 crc kubenswrapper[4837]: I1001 09:14:16.597074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"3f958bc4-38ae-49d4-9a51-205ce949a3b8","Type":"ContainerDied","Data":"a4c5432f9fc74db9bdff070975158e6a095ff2ba0b3b32b32b90127446998f81"} Oct 01 09:14:19 crc kubenswrapper[4837]: I1001 09:14:19.634348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"3f958bc4-38ae-49d4-9a51-205ce949a3b8","Type":"ContainerStarted","Data":"22c203e6b4ca602d1598798a1802d697c6069149de153ddfa9d2b62c45dfa6ea"} Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.079210 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.079531 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.079572 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.080268 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.080324 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" gracePeriod=600 Oct 01 09:14:23 crc kubenswrapper[4837]: E1001 09:14:23.212423 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.690818 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerStarted","Data":"344cc67d25cfd9b1eaf22496e66e077aabf271b71c6f82e1b4c24c5daa828fb3"} Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.694094 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"3f958bc4-38ae-49d4-9a51-205ce949a3b8","Type":"ContainerStarted","Data":"c6ecafe07f96adfdc11b7db32cf3ed4ab25370984034b087dffaeddaf3a873b3"} Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.694346 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.696913 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" exitCode=0 Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.697054 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5"} Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.697235 4837 scope.go:117] "RemoveContainer" containerID="5e2ac5784e622e9c7f219220022a21527aca3b3bfed1f16b99d50215bb5cf810" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.698102 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:14:23 crc kubenswrapper[4837]: E1001 09:14:23.698519 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.701449 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 01 09:14:23 crc kubenswrapper[4837]: I1001 09:14:23.738966 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.593175983 podStartE2EDuration="23.738940822s" podCreationTimestamp="2025-10-01 09:14:00 +0000 UTC" firstStartedPulling="2025-10-01 09:14:01.495823008 +0000 UTC m=+7698.337430463" lastFinishedPulling="2025-10-01 09:14:18.641587847 +0000 UTC m=+7715.483195302" observedRunningTime="2025-10-01 09:14:23.731925188 +0000 UTC m=+7720.573532673" watchObservedRunningTime="2025-10-01 09:14:23.738940822 +0000 UTC m=+7720.580548287" Oct 01 09:14:28 crc kubenswrapper[4837]: I1001 09:14:28.770294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerStarted","Data":"59289f4b3268ab17bb5df51976880b1201a133ba22a876bbc6affb321988cfa3"} Oct 01 09:14:30 crc kubenswrapper[4837]: I1001 09:14:30.050281 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vzspd"] Oct 01 09:14:30 crc kubenswrapper[4837]: I1001 09:14:30.060405 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vzspd"] Oct 01 09:14:31 crc kubenswrapper[4837]: I1001 09:14:31.833286 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5a7970a-45f3-48f2-adec-2a19833c2fe5" path="/var/lib/kubelet/pods/d5a7970a-45f3-48f2-adec-2a19833c2fe5/volumes" Oct 01 09:14:32 crc kubenswrapper[4837]: I1001 09:14:32.818279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerStarted","Data":"86afeb216c3e30b0bb8c0b8b98aa60334105de285eb7488e2bf9d125f83b45d9"} Oct 01 09:14:32 crc kubenswrapper[4837]: I1001 09:14:32.856717 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.935009656 podStartE2EDuration="33.856666356s" podCreationTimestamp="2025-10-01 09:13:59 +0000 UTC" firstStartedPulling="2025-10-01 09:14:02.362886341 +0000 UTC m=+7699.204493796" lastFinishedPulling="2025-10-01 09:14:32.284543011 +0000 UTC m=+7729.126150496" observedRunningTime="2025-10-01 09:14:32.849277203 +0000 UTC m=+7729.690884658" watchObservedRunningTime="2025-10-01 09:14:32.856666356 +0000 UTC m=+7729.698273821" Oct 01 09:14:36 crc kubenswrapper[4837]: I1001 09:14:36.723829 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:36 crc kubenswrapper[4837]: I1001 09:14:36.816836 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:14:36 crc kubenswrapper[4837]: E1001 09:14:36.817312 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.252625 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.258801 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.261951 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.284649 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.304260 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384117 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384163 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhl92\" (UniqueName: \"kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384223 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384258 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384289 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384322 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.384340 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.486536 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhl92\" (UniqueName: \"kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.486850 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.486980 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487096 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487235 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487345 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487919 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487551 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.487384 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.492513 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.492559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.493539 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.499991 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.503195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhl92\" (UniqueName: \"kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92\") pod \"ceilometer-0\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " pod="openstack/ceilometer-0" Oct 01 09:14:40 crc kubenswrapper[4837]: I1001 09:14:40.592395 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:14:41 crc kubenswrapper[4837]: I1001 09:14:41.107583 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:14:41 crc kubenswrapper[4837]: I1001 09:14:41.930203 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerStarted","Data":"b097cc91581440970152dccf14bdb79b406739ae127a29d91a29aafeffb7290a"} Oct 01 09:14:46 crc kubenswrapper[4837]: I1001 09:14:46.723939 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:46 crc kubenswrapper[4837]: I1001 09:14:46.730514 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:46 crc kubenswrapper[4837]: I1001 09:14:46.995971 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerStarted","Data":"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3"} Oct 01 09:14:46 crc kubenswrapper[4837]: I1001 09:14:46.996904 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:47 crc kubenswrapper[4837]: I1001 09:14:47.816580 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:14:47 crc kubenswrapper[4837]: E1001 09:14:47.817376 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.017028 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerStarted","Data":"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612"} Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.781513 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.782329 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" containerName="openstackclient" containerID="cri-o://ac61d35d9557a9ab47274dd971284105785d08d9b32d00dd3da8098fd66db8da" gracePeriod=2 Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.792206 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.817314 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:48 crc kubenswrapper[4837]: E1001 09:14:48.817917 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" containerName="openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.817931 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" containerName="openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.818131 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" containerName="openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.818789 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.842201 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.860292 4837 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" podUID="87d45b64-bd5f-4828-900a-0077a3f6b59f" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.888401 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.888444 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config-secret\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.888513 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.888551 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwjv\" (UniqueName: \"kubernetes.io/projected/87d45b64-bd5f-4828-900a-0077a3f6b59f-kube-api-access-blwjv\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.991407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.991526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config-secret\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.991621 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.991663 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwjv\" (UniqueName: \"kubernetes.io/projected/87d45b64-bd5f-4828-900a-0077a3f6b59f-kube-api-access-blwjv\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.992980 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:48 crc kubenswrapper[4837]: I1001 09:14:48.997970 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-openstack-config-secret\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:49 crc kubenswrapper[4837]: I1001 09:14:49.011865 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87d45b64-bd5f-4828-900a-0077a3f6b59f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:49 crc kubenswrapper[4837]: I1001 09:14:49.029518 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwjv\" (UniqueName: \"kubernetes.io/projected/87d45b64-bd5f-4828-900a-0077a3f6b59f-kube-api-access-blwjv\") pod \"openstackclient\" (UID: \"87d45b64-bd5f-4828-900a-0077a3f6b59f\") " pod="openstack/openstackclient" Oct 01 09:14:49 crc kubenswrapper[4837]: I1001 09:14:49.035703 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerStarted","Data":"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80"} Oct 01 09:14:49 crc kubenswrapper[4837]: I1001 09:14:49.164869 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:49 crc kubenswrapper[4837]: I1001 09:14:49.728289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.047553 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"87d45b64-bd5f-4828-900a-0077a3f6b59f","Type":"ContainerStarted","Data":"bba9738f8d2d5d195aac56309234b07db75c44a5c3cc58542f2626dde0057162"} Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.047910 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"87d45b64-bd5f-4828-900a-0077a3f6b59f","Type":"ContainerStarted","Data":"94344b65cc993cf342f3e46e8a2db61af7106817e6e047bf4c6c151e9d75b7d0"} Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.071260 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.071237482 podStartE2EDuration="2.071237482s" podCreationTimestamp="2025-10-01 09:14:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:14:50.064531745 +0000 UTC m=+7746.906139190" watchObservedRunningTime="2025-10-01 09:14:50.071237482 +0000 UTC m=+7746.912844947" Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.188999 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.189279 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="prometheus" containerID="cri-o://344cc67d25cfd9b1eaf22496e66e077aabf271b71c6f82e1b4c24c5daa828fb3" gracePeriod=600 Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.191264 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="thanos-sidecar" containerID="cri-o://86afeb216c3e30b0bb8c0b8b98aa60334105de285eb7488e2bf9d125f83b45d9" gracePeriod=600 Oct 01 09:14:50 crc kubenswrapper[4837]: I1001 09:14:50.191375 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="config-reloader" containerID="cri-o://59289f4b3268ab17bb5df51976880b1201a133ba22a876bbc6affb321988cfa3" gracePeriod=600 Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.076194 4837 generic.go:334] "Generic (PLEG): container finished" podID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" containerID="ac61d35d9557a9ab47274dd971284105785d08d9b32d00dd3da8098fd66db8da" exitCode=137 Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087044 4837 generic.go:334] "Generic (PLEG): container finished" podID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerID="86afeb216c3e30b0bb8c0b8b98aa60334105de285eb7488e2bf9d125f83b45d9" exitCode=0 Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087069 4837 generic.go:334] "Generic (PLEG): container finished" podID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerID="59289f4b3268ab17bb5df51976880b1201a133ba22a876bbc6affb321988cfa3" exitCode=0 Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087079 4837 generic.go:334] "Generic (PLEG): container finished" podID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerID="344cc67d25cfd9b1eaf22496e66e077aabf271b71c6f82e1b4c24c5daa828fb3" exitCode=0 Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087117 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerDied","Data":"86afeb216c3e30b0bb8c0b8b98aa60334105de285eb7488e2bf9d125f83b45d9"} Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087164 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerDied","Data":"59289f4b3268ab17bb5df51976880b1201a133ba22a876bbc6affb321988cfa3"} Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.087179 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerDied","Data":"344cc67d25cfd9b1eaf22496e66e077aabf271b71c6f82e1b4c24c5daa828fb3"} Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.210168 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.215095 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370488 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370554 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370594 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle\") pod \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370712 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lp9n\" (UniqueName: \"kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n\") pod \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370756 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret\") pod \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370862 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config\") pod \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\" (UID: \"dd1784a7-4be8-4517-a06e-2b3209cd67b8\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370898 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.370930 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.371039 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.371069 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpk2w\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w\") pod \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\" (UID: \"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e\") " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.376055 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.376928 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out" (OuterVolumeSpecName: "config-out") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.386828 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config" (OuterVolumeSpecName: "config") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.387047 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w" (OuterVolumeSpecName: "kube-api-access-hpk2w") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "kube-api-access-hpk2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.387170 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.399799 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n" (OuterVolumeSpecName: "kube-api-access-9lp9n") pod "dd1784a7-4be8-4517-a06e-2b3209cd67b8" (UID: "dd1784a7-4be8-4517-a06e-2b3209cd67b8"). InnerVolumeSpecName "kube-api-access-9lp9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.401235 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.417966 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "pvc-26092500-1c26-427b-9bfd-68974b141be7". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.426568 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config" (OuterVolumeSpecName: "web-config") pod "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" (UID: "8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.439754 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd1784a7-4be8-4517-a06e-2b3209cd67b8" (UID: "dd1784a7-4be8-4517-a06e-2b3209cd67b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.440080 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dd1784a7-4be8-4517-a06e-2b3209cd67b8" (UID: "dd1784a7-4be8-4517-a06e-2b3209cd67b8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474235 4837 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474315 4837 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474346 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") on node \"crc\" " Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474357 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpk2w\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-kube-api-access-hpk2w\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474369 4837 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-web-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474378 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474386 4837 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474394 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474404 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lp9n\" (UniqueName: \"kubernetes.io/projected/dd1784a7-4be8-4517-a06e-2b3209cd67b8-kube-api-access-9lp9n\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474412 4837 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e-config-out\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.474420 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.493763 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dd1784a7-4be8-4517-a06e-2b3209cd67b8" (UID: "dd1784a7-4be8-4517-a06e-2b3209cd67b8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.502939 4837 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.503119 4837 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-26092500-1c26-427b-9bfd-68974b141be7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7") on node "crc" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.576663 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dd1784a7-4be8-4517-a06e-2b3209cd67b8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.577213 4837 reconciler_common.go:293] "Volume detached for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") on node \"crc\" DevicePath \"\"" Oct 01 09:14:51 crc kubenswrapper[4837]: I1001 09:14:51.829327 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1784a7-4be8-4517-a06e-2b3209cd67b8" path="/var/lib/kubelet/pods/dd1784a7-4be8-4517-a06e-2b3209cd67b8/volumes" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.099239 4837 scope.go:117] "RemoveContainer" containerID="ac61d35d9557a9ab47274dd971284105785d08d9b32d00dd3da8098fd66db8da" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.099375 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.107352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e","Type":"ContainerDied","Data":"6b4feaaf42665f3634469d84bd1c4d1859be92e640045b744ab91e9879abb832"} Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.107446 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.114095 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerStarted","Data":"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30"} Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.115169 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.132241 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.135727 4837 scope.go:117] "RemoveContainer" containerID="86afeb216c3e30b0bb8c0b8b98aa60334105de285eb7488e2bf9d125f83b45d9" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.165048 4837 scope.go:117] "RemoveContainer" containerID="59289f4b3268ab17bb5df51976880b1201a133ba22a876bbc6affb321988cfa3" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.186901 4837 scope.go:117] "RemoveContainer" containerID="344cc67d25cfd9b1eaf22496e66e077aabf271b71c6f82e1b4c24c5daa828fb3" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.193197 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.233759 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:52 crc kubenswrapper[4837]: E1001 09:14:52.234212 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="thanos-sidecar" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234227 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="thanos-sidecar" Oct 01 09:14:52 crc kubenswrapper[4837]: E1001 09:14:52.234251 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="prometheus" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234258 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="prometheus" Oct 01 09:14:52 crc kubenswrapper[4837]: E1001 09:14:52.234272 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="config-reloader" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234278 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="config-reloader" Oct 01 09:14:52 crc kubenswrapper[4837]: E1001 09:14:52.234287 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="init-config-reloader" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234293 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="init-config-reloader" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234498 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="prometheus" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234518 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="config-reloader" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234528 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" containerName="thanos-sidecar" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.234744 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.910149949 podStartE2EDuration="12.234727255s" podCreationTimestamp="2025-10-01 09:14:40 +0000 UTC" firstStartedPulling="2025-10-01 09:14:41.103171965 +0000 UTC m=+7737.944779450" lastFinishedPulling="2025-10-01 09:14:51.427749301 +0000 UTC m=+7748.269356756" observedRunningTime="2025-10-01 09:14:52.227275321 +0000 UTC m=+7749.068882776" watchObservedRunningTime="2025-10-01 09:14:52.234727255 +0000 UTC m=+7749.076334710" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.236257 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.246275 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-g867p" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.246466 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.246623 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.246849 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.246855 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.247540 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.266069 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.267051 4837 scope.go:117] "RemoveContainer" containerID="349d0f74f3b69457bbf8cc775c808dcc432384da3de7ffe998cf510a4318c278" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.289947 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400794 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400846 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbnv5\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-kube-api-access-lbnv5\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400909 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dfb858f6-b1bb-41ee-a7be-ecef06894512-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400930 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400982 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.400998 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dfb858f6-b1bb-41ee-a7be-ecef06894512-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.401017 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.401041 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.401068 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.401092 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504821 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504873 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dfb858f6-b1bb-41ee-a7be-ecef06894512-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504902 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504962 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504979 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dfb858f6-b1bb-41ee-a7be-ecef06894512-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.504996 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.505022 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.505050 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.505072 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.505139 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.505162 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbnv5\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-kube-api-access-lbnv5\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.506428 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dfb858f6-b1bb-41ee-a7be-ecef06894512-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.510246 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.510888 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.511256 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.511291 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5a32693e0b704d1d24b4f326abcdde6b5965d6dd0bf58ad7942285c2e3caa226/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.512342 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.512943 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.514180 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dfb858f6-b1bb-41ee-a7be-ecef06894512-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.515053 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.515125 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.527027 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbnv5\" (UniqueName: \"kubernetes.io/projected/dfb858f6-b1bb-41ee-a7be-ecef06894512-kube-api-access-lbnv5\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.535714 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dfb858f6-b1bb-41ee-a7be-ecef06894512-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.570538 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-26092500-1c26-427b-9bfd-68974b141be7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-26092500-1c26-427b-9bfd-68974b141be7\") pod \"prometheus-metric-storage-0\" (UID: \"dfb858f6-b1bb-41ee-a7be-ecef06894512\") " pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:52 crc kubenswrapper[4837]: I1001 09:14:52.602612 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 01 09:14:53 crc kubenswrapper[4837]: I1001 09:14:53.106380 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 01 09:14:53 crc kubenswrapper[4837]: I1001 09:14:53.842826 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e" path="/var/lib/kubelet/pods/8e4d4543-2cf5-4bd2-8b26-57ebefeb0f0e/volumes" Oct 01 09:14:54 crc kubenswrapper[4837]: I1001 09:14:54.152992 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerStarted","Data":"0d1a8699417d8efc59bc6b97b3cf080f3d7903f52d380e78dfcb902270b3b6a9"} Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.412299 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-26rgd"] Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.414071 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.423958 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-26rgd"] Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.505422 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9bm7\" (UniqueName: \"kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7\") pod \"aodh-db-create-26rgd\" (UID: \"890300f8-e949-4add-9103-3ddab9b0de07\") " pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.608865 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9bm7\" (UniqueName: \"kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7\") pod \"aodh-db-create-26rgd\" (UID: \"890300f8-e949-4add-9103-3ddab9b0de07\") " pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:56 crc kubenswrapper[4837]: I1001 09:14:56.748783 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9bm7\" (UniqueName: \"kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7\") pod \"aodh-db-create-26rgd\" (UID: \"890300f8-e949-4add-9103-3ddab9b0de07\") " pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:57 crc kubenswrapper[4837]: I1001 09:14:57.038019 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:57 crc kubenswrapper[4837]: I1001 09:14:57.578496 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-26rgd"] Oct 01 09:14:58 crc kubenswrapper[4837]: I1001 09:14:58.212769 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerStarted","Data":"36fb6ba121390c7891585e0f635204b2062933c3db564070778a5c348567be52"} Oct 01 09:14:58 crc kubenswrapper[4837]: I1001 09:14:58.214582 4837 generic.go:334] "Generic (PLEG): container finished" podID="890300f8-e949-4add-9103-3ddab9b0de07" containerID="36e8a3c24a1ff9c9c4225b06a40a4cea653c78ee753ae9cded7d0886645d48b8" exitCode=0 Oct 01 09:14:58 crc kubenswrapper[4837]: I1001 09:14:58.214626 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-26rgd" event={"ID":"890300f8-e949-4add-9103-3ddab9b0de07","Type":"ContainerDied","Data":"36e8a3c24a1ff9c9c4225b06a40a4cea653c78ee753ae9cded7d0886645d48b8"} Oct 01 09:14:58 crc kubenswrapper[4837]: I1001 09:14:58.214653 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-26rgd" event={"ID":"890300f8-e949-4add-9103-3ddab9b0de07","Type":"ContainerStarted","Data":"9a1be73e87812d1adadb4b9b391a69b5f9cd59884a0d295de9fcbabfd1f2616a"} Oct 01 09:14:59 crc kubenswrapper[4837]: I1001 09:14:59.705707 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-26rgd" Oct 01 09:14:59 crc kubenswrapper[4837]: I1001 09:14:59.787388 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9bm7\" (UniqueName: \"kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7\") pod \"890300f8-e949-4add-9103-3ddab9b0de07\" (UID: \"890300f8-e949-4add-9103-3ddab9b0de07\") " Oct 01 09:14:59 crc kubenswrapper[4837]: I1001 09:14:59.797128 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7" (OuterVolumeSpecName: "kube-api-access-l9bm7") pod "890300f8-e949-4add-9103-3ddab9b0de07" (UID: "890300f8-e949-4add-9103-3ddab9b0de07"). InnerVolumeSpecName "kube-api-access-l9bm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:14:59 crc kubenswrapper[4837]: I1001 09:14:59.890519 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9bm7\" (UniqueName: \"kubernetes.io/projected/890300f8-e949-4add-9103-3ddab9b0de07-kube-api-access-l9bm7\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.177522 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9"] Oct 01 09:15:00 crc kubenswrapper[4837]: E1001 09:15:00.178434 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890300f8-e949-4add-9103-3ddab9b0de07" containerName="mariadb-database-create" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.178493 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="890300f8-e949-4add-9103-3ddab9b0de07" containerName="mariadb-database-create" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.179004 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="890300f8-e949-4add-9103-3ddab9b0de07" containerName="mariadb-database-create" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.180757 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.183994 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.185231 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.194122 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9"] Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.237033 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-26rgd" event={"ID":"890300f8-e949-4add-9103-3ddab9b0de07","Type":"ContainerDied","Data":"9a1be73e87812d1adadb4b9b391a69b5f9cd59884a0d295de9fcbabfd1f2616a"} Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.237406 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a1be73e87812d1adadb4b9b391a69b5f9cd59884a0d295de9fcbabfd1f2616a" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.237099 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-26rgd" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.300494 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9z9n\" (UniqueName: \"kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.300656 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.300810 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.403373 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9z9n\" (UniqueName: \"kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.403477 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.403567 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.404755 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.417843 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.425009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9z9n\" (UniqueName: \"kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n\") pod \"collect-profiles-29321835-rp4m9\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:00 crc kubenswrapper[4837]: I1001 09:15:00.510630 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:01 crc kubenswrapper[4837]: I1001 09:15:01.010625 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9"] Oct 01 09:15:01 crc kubenswrapper[4837]: W1001 09:15:01.011103 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65feda55_7457_461d_abd9_a80d76d421c3.slice/crio-b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad WatchSource:0}: Error finding container b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad: Status 404 returned error can't find the container with id b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad Oct 01 09:15:01 crc kubenswrapper[4837]: I1001 09:15:01.250522 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" event={"ID":"65feda55-7457-461d-abd9-a80d76d421c3","Type":"ContainerStarted","Data":"b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad"} Oct 01 09:15:01 crc kubenswrapper[4837]: I1001 09:15:01.817468 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:15:01 crc kubenswrapper[4837]: E1001 09:15:01.818537 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:15:02 crc kubenswrapper[4837]: I1001 09:15:02.268040 4837 generic.go:334] "Generic (PLEG): container finished" podID="65feda55-7457-461d-abd9-a80d76d421c3" containerID="a4462a7845d32476f3a0f3230099099a685b1d0e8f7bdde77d5c993ccfc5190c" exitCode=0 Oct 01 09:15:02 crc kubenswrapper[4837]: I1001 09:15:02.268096 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" event={"ID":"65feda55-7457-461d-abd9-a80d76d421c3","Type":"ContainerDied","Data":"a4462a7845d32476f3a0f3230099099a685b1d0e8f7bdde77d5c993ccfc5190c"} Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.725118 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.784053 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9z9n\" (UniqueName: \"kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n\") pod \"65feda55-7457-461d-abd9-a80d76d421c3\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.784480 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume\") pod \"65feda55-7457-461d-abd9-a80d76d421c3\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.784783 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume\") pod \"65feda55-7457-461d-abd9-a80d76d421c3\" (UID: \"65feda55-7457-461d-abd9-a80d76d421c3\") " Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.785208 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume" (OuterVolumeSpecName: "config-volume") pod "65feda55-7457-461d-abd9-a80d76d421c3" (UID: "65feda55-7457-461d-abd9-a80d76d421c3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.785894 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65feda55-7457-461d-abd9-a80d76d421c3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.791612 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "65feda55-7457-461d-abd9-a80d76d421c3" (UID: "65feda55-7457-461d-abd9-a80d76d421c3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.792027 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n" (OuterVolumeSpecName: "kube-api-access-r9z9n") pod "65feda55-7457-461d-abd9-a80d76d421c3" (UID: "65feda55-7457-461d-abd9-a80d76d421c3"). InnerVolumeSpecName "kube-api-access-r9z9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.888219 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9z9n\" (UniqueName: \"kubernetes.io/projected/65feda55-7457-461d-abd9-a80d76d421c3-kube-api-access-r9z9n\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:03 crc kubenswrapper[4837]: I1001 09:15:03.888269 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65feda55-7457-461d-abd9-a80d76d421c3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:04 crc kubenswrapper[4837]: I1001 09:15:04.292941 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" event={"ID":"65feda55-7457-461d-abd9-a80d76d421c3","Type":"ContainerDied","Data":"b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad"} Oct 01 09:15:04 crc kubenswrapper[4837]: I1001 09:15:04.292999 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b07bb3914ddbcb0b1ee854401aae0b5b803d618e6d177c73292adcc6b67a75ad" Oct 01 09:15:04 crc kubenswrapper[4837]: I1001 09:15:04.293052 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9" Oct 01 09:15:04 crc kubenswrapper[4837]: I1001 09:15:04.842918 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7"] Oct 01 09:15:04 crc kubenswrapper[4837]: I1001 09:15:04.859061 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321790-fpbs7"] Oct 01 09:15:05 crc kubenswrapper[4837]: I1001 09:15:05.845804 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30331db9-fb15-4dc8-a777-45201a20608f" path="/var/lib/kubelet/pods/30331db9-fb15-4dc8-a777-45201a20608f/volumes" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.566785 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-79c0-account-create-gvrk7"] Oct 01 09:15:06 crc kubenswrapper[4837]: E1001 09:15:06.568149 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65feda55-7457-461d-abd9-a80d76d421c3" containerName="collect-profiles" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.568193 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="65feda55-7457-461d-abd9-a80d76d421c3" containerName="collect-profiles" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.568640 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="65feda55-7457-461d-abd9-a80d76d421c3" containerName="collect-profiles" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.580059 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-79c0-account-create-gvrk7"] Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.580942 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.584948 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.655094 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8x9\" (UniqueName: \"kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9\") pod \"aodh-79c0-account-create-gvrk7\" (UID: \"2b035865-59b0-46f5-ae64-877c1ce99e05\") " pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.756759 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8x9\" (UniqueName: \"kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9\") pod \"aodh-79c0-account-create-gvrk7\" (UID: \"2b035865-59b0-46f5-ae64-877c1ce99e05\") " pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.783001 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8x9\" (UniqueName: \"kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9\") pod \"aodh-79c0-account-create-gvrk7\" (UID: \"2b035865-59b0-46f5-ae64-877c1ce99e05\") " pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:06 crc kubenswrapper[4837]: I1001 09:15:06.965070 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:07 crc kubenswrapper[4837]: I1001 09:15:07.363741 4837 generic.go:334] "Generic (PLEG): container finished" podID="dfb858f6-b1bb-41ee-a7be-ecef06894512" containerID="36fb6ba121390c7891585e0f635204b2062933c3db564070778a5c348567be52" exitCode=0 Oct 01 09:15:07 crc kubenswrapper[4837]: I1001 09:15:07.363851 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerDied","Data":"36fb6ba121390c7891585e0f635204b2062933c3db564070778a5c348567be52"} Oct 01 09:15:07 crc kubenswrapper[4837]: I1001 09:15:07.485377 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-79c0-account-create-gvrk7"] Oct 01 09:15:07 crc kubenswrapper[4837]: W1001 09:15:07.490200 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b035865_59b0_46f5_ae64_877c1ce99e05.slice/crio-c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f WatchSource:0}: Error finding container c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f: Status 404 returned error can't find the container with id c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f Oct 01 09:15:08 crc kubenswrapper[4837]: I1001 09:15:08.378922 4837 generic.go:334] "Generic (PLEG): container finished" podID="2b035865-59b0-46f5-ae64-877c1ce99e05" containerID="d1e73b6a3db5f5773c31234ea6dfd7fecc50495c288ddfd8f49523bb89237f14" exitCode=0 Oct 01 09:15:08 crc kubenswrapper[4837]: I1001 09:15:08.379157 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-79c0-account-create-gvrk7" event={"ID":"2b035865-59b0-46f5-ae64-877c1ce99e05","Type":"ContainerDied","Data":"d1e73b6a3db5f5773c31234ea6dfd7fecc50495c288ddfd8f49523bb89237f14"} Oct 01 09:15:08 crc kubenswrapper[4837]: I1001 09:15:08.379318 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-79c0-account-create-gvrk7" event={"ID":"2b035865-59b0-46f5-ae64-877c1ce99e05","Type":"ContainerStarted","Data":"c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f"} Oct 01 09:15:08 crc kubenswrapper[4837]: I1001 09:15:08.381968 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerStarted","Data":"3a51c314f729c270da03dbcdad91d28a3cefc138358c566be1c84618b3cb7730"} Oct 01 09:15:09 crc kubenswrapper[4837]: I1001 09:15:09.926967 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.036977 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8x9\" (UniqueName: \"kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9\") pod \"2b035865-59b0-46f5-ae64-877c1ce99e05\" (UID: \"2b035865-59b0-46f5-ae64-877c1ce99e05\") " Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.042359 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9" (OuterVolumeSpecName: "kube-api-access-qp8x9") pod "2b035865-59b0-46f5-ae64-877c1ce99e05" (UID: "2b035865-59b0-46f5-ae64-877c1ce99e05"). InnerVolumeSpecName "kube-api-access-qp8x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.139799 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8x9\" (UniqueName: \"kubernetes.io/projected/2b035865-59b0-46f5-ae64-877c1ce99e05-kube-api-access-qp8x9\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.428727 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-79c0-account-create-gvrk7" event={"ID":"2b035865-59b0-46f5-ae64-877c1ce99e05","Type":"ContainerDied","Data":"c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f"} Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.429597 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6281b8529ebd573aa1fa4c990617325b199b8d40ea19c4ac7fdd5bd92e75f1f" Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.428980 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-79c0-account-create-gvrk7" Oct 01 09:15:10 crc kubenswrapper[4837]: I1001 09:15:10.606770 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 09:15:11 crc kubenswrapper[4837]: I1001 09:15:11.626382 4837 scope.go:117] "RemoveContainer" containerID="00990d322d65cab6f5c21f980d551e418ccf62d5747dca9d385e50863ce1f874" Oct 01 09:15:11 crc kubenswrapper[4837]: I1001 09:15:11.670564 4837 scope.go:117] "RemoveContainer" containerID="daafde44527af8306a7caeff258126a05924fc5ce3cdf085d25aee55f3bdf7b4" Oct 01 09:15:11 crc kubenswrapper[4837]: I1001 09:15:11.700277 4837 scope.go:117] "RemoveContainer" containerID="d18644e064fa922c7cfb2d3a470fa39c19d2d3738633b7f5e9788fe24bfcbde3" Oct 01 09:15:11 crc kubenswrapper[4837]: I1001 09:15:11.734414 4837 scope.go:117] "RemoveContainer" containerID="b89b918c7b46211af038bbb7d58bf08e5181f26962ca0b4b0a461a72f4a0a8ac" Oct 01 09:15:11 crc kubenswrapper[4837]: I1001 09:15:11.766945 4837 scope.go:117] "RemoveContainer" containerID="d7070aa49b6860632cba07b25ba3b2170cd02587dce8b98332874bcbfc39ddd6" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.022998 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-6fg5l"] Oct 01 09:15:12 crc kubenswrapper[4837]: E1001 09:15:12.023478 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b035865-59b0-46f5-ae64-877c1ce99e05" containerName="mariadb-account-create" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.023498 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b035865-59b0-46f5-ae64-877c1ce99e05" containerName="mariadb-account-create" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.023759 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b035865-59b0-46f5-ae64-877c1ce99e05" containerName="mariadb-account-create" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.024537 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.026449 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.026684 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.026942 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-x82kr" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.049670 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6fg5l"] Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.103896 4837 scope.go:117] "RemoveContainer" containerID="1a707fc0da32a6a48304a5601857ee0a448895c615ec0b15a4454cfdfd7e442c" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.185294 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zljzj\" (UniqueName: \"kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.185423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.185564 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.185629 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.287372 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.287430 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.287458 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.287530 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zljzj\" (UniqueName: \"kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.292567 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.292592 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.292992 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.301994 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zljzj\" (UniqueName: \"kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj\") pod \"aodh-db-sync-6fg5l\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.345543 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:12 crc kubenswrapper[4837]: I1001 09:15:12.823153 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6fg5l"] Oct 01 09:15:12 crc kubenswrapper[4837]: W1001 09:15:12.824823 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6fa401e_5355_41db_8257_e0f97eab28de.slice/crio-c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21 WatchSource:0}: Error finding container c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21: Status 404 returned error can't find the container with id c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21 Oct 01 09:15:13 crc kubenswrapper[4837]: I1001 09:15:13.483521 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6fg5l" event={"ID":"c6fa401e-5355-41db-8257-e0f97eab28de","Type":"ContainerStarted","Data":"c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21"} Oct 01 09:15:14 crc kubenswrapper[4837]: I1001 09:15:14.523815 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerStarted","Data":"a5d371256c54f95d8a6a40c05dbc60931014da97b3307ff3a0c0488873c72d29"} Oct 01 09:15:14 crc kubenswrapper[4837]: I1001 09:15:14.524266 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dfb858f6-b1bb-41ee-a7be-ecef06894512","Type":"ContainerStarted","Data":"e7261c6ee41dd912bf3d9a82bb05f852cecad5b720a419218f71c075ff121708"} Oct 01 09:15:14 crc kubenswrapper[4837]: I1001 09:15:14.573070 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=22.573049138000002 podStartE2EDuration="22.573049138s" podCreationTimestamp="2025-10-01 09:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:15:14.553738501 +0000 UTC m=+7771.395345976" watchObservedRunningTime="2025-10-01 09:15:14.573049138 +0000 UTC m=+7771.414656593" Oct 01 09:15:15 crc kubenswrapper[4837]: I1001 09:15:15.303293 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:15 crc kubenswrapper[4837]: I1001 09:15:15.303515 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="30220513-03a1-44c1-b372-b878b9701a6e" containerName="kube-state-metrics" containerID="cri-o://fbc28bf12bdab8f685750f34986cdd0c7fd27df5a85c92937f42f44fbc0a414c" gracePeriod=30 Oct 01 09:15:15 crc kubenswrapper[4837]: I1001 09:15:15.536985 4837 generic.go:334] "Generic (PLEG): container finished" podID="30220513-03a1-44c1-b372-b878b9701a6e" containerID="fbc28bf12bdab8f685750f34986cdd0c7fd27df5a85c92937f42f44fbc0a414c" exitCode=2 Oct 01 09:15:15 crc kubenswrapper[4837]: I1001 09:15:15.537068 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30220513-03a1-44c1-b372-b878b9701a6e","Type":"ContainerDied","Data":"fbc28bf12bdab8f685750f34986cdd0c7fd27df5a85c92937f42f44fbc0a414c"} Oct 01 09:15:15 crc kubenswrapper[4837]: I1001 09:15:15.816184 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:15:15 crc kubenswrapper[4837]: E1001 09:15:15.816668 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:15:16 crc kubenswrapper[4837]: I1001 09:15:16.998107 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:16 crc kubenswrapper[4837]: I1001 09:15:16.998416 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-central-agent" containerID="cri-o://e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3" gracePeriod=30 Oct 01 09:15:16 crc kubenswrapper[4837]: I1001 09:15:16.998494 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="proxy-httpd" containerID="cri-o://08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30" gracePeriod=30 Oct 01 09:15:16 crc kubenswrapper[4837]: I1001 09:15:16.998658 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-notification-agent" containerID="cri-o://88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612" gracePeriod=30 Oct 01 09:15:16 crc kubenswrapper[4837]: I1001 09:15:16.998708 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="sg-core" containerID="cri-o://14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80" gracePeriod=30 Oct 01 09:15:17 crc kubenswrapper[4837]: I1001 09:15:17.565726 4837 generic.go:334] "Generic (PLEG): container finished" podID="4c349645-8794-4915-8642-32039878ff40" containerID="08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30" exitCode=0 Oct 01 09:15:17 crc kubenswrapper[4837]: I1001 09:15:17.565976 4837 generic.go:334] "Generic (PLEG): container finished" podID="4c349645-8794-4915-8642-32039878ff40" containerID="14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80" exitCode=2 Oct 01 09:15:17 crc kubenswrapper[4837]: I1001 09:15:17.565746 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerDied","Data":"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30"} Oct 01 09:15:17 crc kubenswrapper[4837]: I1001 09:15:17.566012 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerDied","Data":"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80"} Oct 01 09:15:17 crc kubenswrapper[4837]: I1001 09:15:17.602728 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.579254 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.591862 4837 generic.go:334] "Generic (PLEG): container finished" podID="4c349645-8794-4915-8642-32039878ff40" containerID="e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3" exitCode=0 Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.591952 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerDied","Data":"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3"} Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.599000 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6fg5l" event={"ID":"c6fa401e-5355-41db-8257-e0f97eab28de","Type":"ContainerStarted","Data":"c213a609b3f454910eaa1fb8ca621dfbec8fd94396bb083b312a832241c19640"} Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.616520 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30220513-03a1-44c1-b372-b878b9701a6e","Type":"ContainerDied","Data":"5d17d62e7bcc3a23de4b5f8cafb630269c6671230701c2ea8b9c1cd842a9559d"} Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.616585 4837 scope.go:117] "RemoveContainer" containerID="fbc28bf12bdab8f685750f34986cdd0c7fd27df5a85c92937f42f44fbc0a414c" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.616730 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.632794 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-6fg5l" podStartSLOduration=2.18181418 podStartE2EDuration="7.632766914s" podCreationTimestamp="2025-10-01 09:15:11 +0000 UTC" firstStartedPulling="2025-10-01 09:15:12.826418784 +0000 UTC m=+7769.668026239" lastFinishedPulling="2025-10-01 09:15:18.277371518 +0000 UTC m=+7775.118978973" observedRunningTime="2025-10-01 09:15:18.622857379 +0000 UTC m=+7775.464464834" watchObservedRunningTime="2025-10-01 09:15:18.632766914 +0000 UTC m=+7775.474374369" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.749825 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6wct\" (UniqueName: \"kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct\") pod \"30220513-03a1-44c1-b372-b878b9701a6e\" (UID: \"30220513-03a1-44c1-b372-b878b9701a6e\") " Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.758915 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct" (OuterVolumeSpecName: "kube-api-access-p6wct") pod "30220513-03a1-44c1-b372-b878b9701a6e" (UID: "30220513-03a1-44c1-b372-b878b9701a6e"). InnerVolumeSpecName "kube-api-access-p6wct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.852566 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6wct\" (UniqueName: \"kubernetes.io/projected/30220513-03a1-44c1-b372-b878b9701a6e-kube-api-access-p6wct\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.975095 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.986179 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.999051 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:18 crc kubenswrapper[4837]: E1001 09:15:18.999583 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30220513-03a1-44c1-b372-b878b9701a6e" containerName="kube-state-metrics" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.999604 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="30220513-03a1-44c1-b372-b878b9701a6e" containerName="kube-state-metrics" Oct 01 09:15:18 crc kubenswrapper[4837]: I1001 09:15:18.999919 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="30220513-03a1-44c1-b372-b878b9701a6e" containerName="kube-state-metrics" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.000799 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.003983 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.004085 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.014185 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.157946 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.158043 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.158073 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.158153 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d77x\" (UniqueName: \"kubernetes.io/projected/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-api-access-7d77x\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.260004 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d77x\" (UniqueName: \"kubernetes.io/projected/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-api-access-7d77x\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.260379 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.260407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.260435 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.265498 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.265566 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.265580 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecafde02-447b-483c-bbe5-218cf0d8f772-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.278197 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d77x\" (UniqueName: \"kubernetes.io/projected/ecafde02-447b-483c-bbe5-218cf0d8f772-kube-api-access-7d77x\") pod \"kube-state-metrics-0\" (UID: \"ecafde02-447b-483c-bbe5-218cf0d8f772\") " pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.323177 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.361758 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465516 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhl92\" (UniqueName: \"kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465560 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465599 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465658 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465727 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.465871 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle\") pod \"4c349645-8794-4915-8642-32039878ff40\" (UID: \"4c349645-8794-4915-8642-32039878ff40\") " Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.467386 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.468543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.470815 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts" (OuterVolumeSpecName: "scripts") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.471038 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92" (OuterVolumeSpecName: "kube-api-access-qhl92") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "kube-api-access-qhl92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.498368 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.560216 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567502 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567531 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhl92\" (UniqueName: \"kubernetes.io/projected/4c349645-8794-4915-8642-32039878ff40-kube-api-access-qhl92\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567541 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567550 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c349645-8794-4915-8642-32039878ff40-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567559 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.567568 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.586995 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data" (OuterVolumeSpecName: "config-data") pod "4c349645-8794-4915-8642-32039878ff40" (UID: "4c349645-8794-4915-8642-32039878ff40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.641845 4837 generic.go:334] "Generic (PLEG): container finished" podID="4c349645-8794-4915-8642-32039878ff40" containerID="88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612" exitCode=0 Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.641896 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerDied","Data":"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612"} Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.641952 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c349645-8794-4915-8642-32039878ff40","Type":"ContainerDied","Data":"b097cc91581440970152dccf14bdb79b406739ae127a29d91a29aafeffb7290a"} Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.641977 4837 scope.go:117] "RemoveContainer" containerID="08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.641908 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.664880 4837 scope.go:117] "RemoveContainer" containerID="14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.674981 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c349645-8794-4915-8642-32039878ff40-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.687030 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.714182 4837 scope.go:117] "RemoveContainer" containerID="88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.716677 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.730621 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.731087 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-notification-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731098 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-notification-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.731118 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-central-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731126 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-central-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.731141 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="sg-core" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731147 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="sg-core" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.731156 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="proxy-httpd" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731163 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="proxy-httpd" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731352 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="sg-core" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731359 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="proxy-httpd" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731377 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-notification-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.731393 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c349645-8794-4915-8642-32039878ff40" containerName="ceilometer-central-agent" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.735683 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.735834 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.742165 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.742877 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.742989 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.805603 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.807160 4837 scope.go:117] "RemoveContainer" containerID="e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3" Oct 01 09:15:19 crc kubenswrapper[4837]: W1001 09:15:19.812205 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecafde02_447b_483c_bbe5_218cf0d8f772.slice/crio-52e571f6eb008dee44d5051d60a674e70e3dc19f0c27277ae73bb0c8d3502478 WatchSource:0}: Error finding container 52e571f6eb008dee44d5051d60a674e70e3dc19f0c27277ae73bb0c8d3502478: Status 404 returned error can't find the container with id 52e571f6eb008dee44d5051d60a674e70e3dc19f0c27277ae73bb0c8d3502478 Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.828163 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30220513-03a1-44c1-b372-b878b9701a6e" path="/var/lib/kubelet/pods/30220513-03a1-44c1-b372-b878b9701a6e/volumes" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.828725 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c349645-8794-4915-8642-32039878ff40" path="/var/lib/kubelet/pods/4c349645-8794-4915-8642-32039878ff40/volumes" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.870152 4837 scope.go:117] "RemoveContainer" containerID="08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.870877 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30\": container with ID starting with 08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30 not found: ID does not exist" containerID="08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.870921 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30"} err="failed to get container status \"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30\": rpc error: code = NotFound desc = could not find container \"08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30\": container with ID starting with 08f1022866f0b903e4ca9ed16181ef200a858b457ab09285a9e9c8ed1bde1c30 not found: ID does not exist" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.870949 4837 scope.go:117] "RemoveContainer" containerID="14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.871373 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80\": container with ID starting with 14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80 not found: ID does not exist" containerID="14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.871431 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80"} err="failed to get container status \"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80\": rpc error: code = NotFound desc = could not find container \"14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80\": container with ID starting with 14312aa76bc89d73fd511a609a4658425fcd35f37606cc05671db49c23908b80 not found: ID does not exist" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.871472 4837 scope.go:117] "RemoveContainer" containerID="88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.871946 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612\": container with ID starting with 88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612 not found: ID does not exist" containerID="88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.871982 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612"} err="failed to get container status \"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612\": rpc error: code = NotFound desc = could not find container \"88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612\": container with ID starting with 88fb717936916c35cd86e03d81497053bbf939fda0bfc0a9a67700229a87c612 not found: ID does not exist" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.872007 4837 scope.go:117] "RemoveContainer" containerID="e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3" Oct 01 09:15:19 crc kubenswrapper[4837]: E1001 09:15:19.872301 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3\": container with ID starting with e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3 not found: ID does not exist" containerID="e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.872335 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3"} err="failed to get container status \"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3\": rpc error: code = NotFound desc = could not find container \"e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3\": container with ID starting with e8c0b9872b579300c761663e0d45cdcc66d8f17cc55a398ffb1028f95eda08c3 not found: ID does not exist" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880397 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880472 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880499 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tj57\" (UniqueName: \"kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880531 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880571 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.880927 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.881031 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.881109 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.983925 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.983994 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984028 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984065 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984107 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984129 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tj57\" (UniqueName: \"kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984160 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984209 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.984681 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:19 crc kubenswrapper[4837]: I1001 09:15:19.985130 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:19.999967 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.001515 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.001851 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tj57\" (UniqueName: \"kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.002066 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.002764 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.003410 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts\") pod \"ceilometer-0\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.102054 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.554650 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:20 crc kubenswrapper[4837]: W1001 09:15:20.565165 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35c38289_b99f_44d9_9909_b97ea4bef6db.slice/crio-a19fdea86454f7fcd5a0f759f546703e320e8bffd78ab48cd8ea8e33db695987 WatchSource:0}: Error finding container a19fdea86454f7fcd5a0f759f546703e320e8bffd78ab48cd8ea8e33db695987: Status 404 returned error can't find the container with id a19fdea86454f7fcd5a0f759f546703e320e8bffd78ab48cd8ea8e33db695987 Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.655410 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerStarted","Data":"a19fdea86454f7fcd5a0f759f546703e320e8bffd78ab48cd8ea8e33db695987"} Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.658503 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecafde02-447b-483c-bbe5-218cf0d8f772","Type":"ContainerStarted","Data":"f41b1b4c6fad77d7d536b47991948927d8f09e232a446a51defb6b74d7263d23"} Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.658527 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecafde02-447b-483c-bbe5-218cf0d8f772","Type":"ContainerStarted","Data":"52e571f6eb008dee44d5051d60a674e70e3dc19f0c27277ae73bb0c8d3502478"} Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.659726 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 09:15:20 crc kubenswrapper[4837]: I1001 09:15:20.689027 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.2982519200000002 podStartE2EDuration="2.689010511s" podCreationTimestamp="2025-10-01 09:15:18 +0000 UTC" firstStartedPulling="2025-10-01 09:15:19.814407031 +0000 UTC m=+7776.656014486" lastFinishedPulling="2025-10-01 09:15:20.205165632 +0000 UTC m=+7777.046773077" observedRunningTime="2025-10-01 09:15:20.681301149 +0000 UTC m=+7777.522908604" watchObservedRunningTime="2025-10-01 09:15:20.689010511 +0000 UTC m=+7777.530617966" Oct 01 09:15:21 crc kubenswrapper[4837]: I1001 09:15:21.673511 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6fg5l" event={"ID":"c6fa401e-5355-41db-8257-e0f97eab28de","Type":"ContainerDied","Data":"c213a609b3f454910eaa1fb8ca621dfbec8fd94396bb083b312a832241c19640"} Oct 01 09:15:21 crc kubenswrapper[4837]: I1001 09:15:21.673636 4837 generic.go:334] "Generic (PLEG): container finished" podID="c6fa401e-5355-41db-8257-e0f97eab28de" containerID="c213a609b3f454910eaa1fb8ca621dfbec8fd94396bb083b312a832241c19640" exitCode=0 Oct 01 09:15:21 crc kubenswrapper[4837]: I1001 09:15:21.676071 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerStarted","Data":"95fb7c5dfa6b87f14162660e1aa6f9d065c932873d8ee2aba70eda3c3370a785"} Oct 01 09:15:21 crc kubenswrapper[4837]: I1001 09:15:21.676135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerStarted","Data":"01fb19339134679dca0f62bfab294b7a1733ecf53e1770705e42a754610aed52"} Oct 01 09:15:22 crc kubenswrapper[4837]: I1001 09:15:22.603709 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 01 09:15:22 crc kubenswrapper[4837]: I1001 09:15:22.611652 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 01 09:15:22 crc kubenswrapper[4837]: I1001 09:15:22.690286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerStarted","Data":"51edb1e34a34ad8c44981bb1d0b3d1a436dc9b4e915aa3af6c8970e82cc36abb"} Oct 01 09:15:22 crc kubenswrapper[4837]: I1001 09:15:22.695825 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.227941 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.383557 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle\") pod \"c6fa401e-5355-41db-8257-e0f97eab28de\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.383611 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data\") pod \"c6fa401e-5355-41db-8257-e0f97eab28de\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.383664 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts\") pod \"c6fa401e-5355-41db-8257-e0f97eab28de\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.383737 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zljzj\" (UniqueName: \"kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj\") pod \"c6fa401e-5355-41db-8257-e0f97eab28de\" (UID: \"c6fa401e-5355-41db-8257-e0f97eab28de\") " Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.394369 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj" (OuterVolumeSpecName: "kube-api-access-zljzj") pod "c6fa401e-5355-41db-8257-e0f97eab28de" (UID: "c6fa401e-5355-41db-8257-e0f97eab28de"). InnerVolumeSpecName "kube-api-access-zljzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.396028 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts" (OuterVolumeSpecName: "scripts") pod "c6fa401e-5355-41db-8257-e0f97eab28de" (UID: "c6fa401e-5355-41db-8257-e0f97eab28de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.413809 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6fa401e-5355-41db-8257-e0f97eab28de" (UID: "c6fa401e-5355-41db-8257-e0f97eab28de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.415374 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data" (OuterVolumeSpecName: "config-data") pod "c6fa401e-5355-41db-8257-e0f97eab28de" (UID: "c6fa401e-5355-41db-8257-e0f97eab28de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.486341 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zljzj\" (UniqueName: \"kubernetes.io/projected/c6fa401e-5355-41db-8257-e0f97eab28de-kube-api-access-zljzj\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.486597 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.486610 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.486680 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6fa401e-5355-41db-8257-e0f97eab28de-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.700264 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6fg5l" event={"ID":"c6fa401e-5355-41db-8257-e0f97eab28de","Type":"ContainerDied","Data":"c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21"} Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.700321 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4e40dba29f93d2534c282cc2ed41512981170a05a791c07153cfeaa8904cb21" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.700286 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6fg5l" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.703398 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerStarted","Data":"b26dda6733cd762ee75c30e5a82684efa12d85fe6e0e2dd6bc717c10ae24bf82"} Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.703449 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:15:23 crc kubenswrapper[4837]: I1001 09:15:23.743027 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8875378889999999 podStartE2EDuration="4.743003208s" podCreationTimestamp="2025-10-01 09:15:19 +0000 UTC" firstStartedPulling="2025-10-01 09:15:20.568111725 +0000 UTC m=+7777.409719180" lastFinishedPulling="2025-10-01 09:15:23.423577054 +0000 UTC m=+7780.265184499" observedRunningTime="2025-10-01 09:15:23.732857196 +0000 UTC m=+7780.574464651" watchObservedRunningTime="2025-10-01 09:15:23.743003208 +0000 UTC m=+7780.584610663" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.415356 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 01 09:15:26 crc kubenswrapper[4837]: E1001 09:15:26.416000 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fa401e-5355-41db-8257-e0f97eab28de" containerName="aodh-db-sync" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.416016 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fa401e-5355-41db-8257-e0f97eab28de" containerName="aodh-db-sync" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.416194 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fa401e-5355-41db-8257-e0f97eab28de" containerName="aodh-db-sync" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.418389 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.419833 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-x82kr" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.421225 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.421426 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.434230 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.546453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.546519 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.546629 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.546798 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggh8t\" (UniqueName: \"kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.649021 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.649079 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.649177 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.649224 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggh8t\" (UniqueName: \"kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.655140 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.658384 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.659258 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.667275 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggh8t\" (UniqueName: \"kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t\") pod \"aodh-0\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.753342 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:15:26 crc kubenswrapper[4837]: I1001 09:15:26.816602 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:15:26 crc kubenswrapper[4837]: E1001 09:15:26.816848 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:15:27 crc kubenswrapper[4837]: W1001 09:15:27.268660 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37190d42_7612_4adc_ae7f_3ce3917b3e16.slice/crio-03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996 WatchSource:0}: Error finding container 03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996: Status 404 returned error can't find the container with id 03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996 Oct 01 09:15:27 crc kubenswrapper[4837]: I1001 09:15:27.269690 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 01 09:15:27 crc kubenswrapper[4837]: I1001 09:15:27.738616 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerStarted","Data":"a8462774cf7a801136982b3a4b5c916ba1c45069d9a6874c59d2fa526584b18c"} Oct 01 09:15:27 crc kubenswrapper[4837]: I1001 09:15:27.738986 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerStarted","Data":"03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.002369 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.003439 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-central-agent" containerID="cri-o://01fb19339134679dca0f62bfab294b7a1733ecf53e1770705e42a754610aed52" gracePeriod=30 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.004001 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="proxy-httpd" containerID="cri-o://b26dda6733cd762ee75c30e5a82684efa12d85fe6e0e2dd6bc717c10ae24bf82" gracePeriod=30 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.004068 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="sg-core" containerID="cri-o://51edb1e34a34ad8c44981bb1d0b3d1a436dc9b4e915aa3af6c8970e82cc36abb" gracePeriod=30 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.004037 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-notification-agent" containerID="cri-o://95fb7c5dfa6b87f14162660e1aa6f9d065c932873d8ee2aba70eda3c3370a785" gracePeriod=30 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.346414 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.539052 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.762150 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerStarted","Data":"91582e7ecb939bb8896ec2ec40a8583209bc646b3e57287e25b9c79db07672e9"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764488 4837 generic.go:334] "Generic (PLEG): container finished" podID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerID="b26dda6733cd762ee75c30e5a82684efa12d85fe6e0e2dd6bc717c10ae24bf82" exitCode=0 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764506 4837 generic.go:334] "Generic (PLEG): container finished" podID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerID="51edb1e34a34ad8c44981bb1d0b3d1a436dc9b4e915aa3af6c8970e82cc36abb" exitCode=2 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764516 4837 generic.go:334] "Generic (PLEG): container finished" podID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerID="95fb7c5dfa6b87f14162660e1aa6f9d065c932873d8ee2aba70eda3c3370a785" exitCode=0 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764523 4837 generic.go:334] "Generic (PLEG): container finished" podID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerID="01fb19339134679dca0f62bfab294b7a1733ecf53e1770705e42a754610aed52" exitCode=0 Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764536 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerDied","Data":"b26dda6733cd762ee75c30e5a82684efa12d85fe6e0e2dd6bc717c10ae24bf82"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764552 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerDied","Data":"51edb1e34a34ad8c44981bb1d0b3d1a436dc9b4e915aa3af6c8970e82cc36abb"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764562 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerDied","Data":"95fb7c5dfa6b87f14162660e1aa6f9d065c932873d8ee2aba70eda3c3370a785"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.764570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerDied","Data":"01fb19339134679dca0f62bfab294b7a1733ecf53e1770705e42a754610aed52"} Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.937923 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.961895 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962018 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962088 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tj57\" (UniqueName: \"kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962135 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962229 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962263 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962396 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle\") pod \"35c38289-b99f-44d9-9909-b97ea4bef6db\" (UID: \"35c38289-b99f-44d9-9909-b97ea4bef6db\") " Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962621 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962639 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962967 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.962990 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/35c38289-b99f-44d9-9909-b97ea4bef6db-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.971714 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57" (OuterVolumeSpecName: "kube-api-access-8tj57") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "kube-api-access-8tj57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:29 crc kubenswrapper[4837]: I1001 09:15:29.971841 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts" (OuterVolumeSpecName: "scripts") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.048198 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.051824 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.064070 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.064101 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.064111 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.064119 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tj57\" (UniqueName: \"kubernetes.io/projected/35c38289-b99f-44d9-9909-b97ea4bef6db-kube-api-access-8tj57\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.068789 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.085902 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data" (OuterVolumeSpecName: "config-data") pod "35c38289-b99f-44d9-9909-b97ea4bef6db" (UID: "35c38289-b99f-44d9-9909-b97ea4bef6db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.165795 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.165833 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/35c38289-b99f-44d9-9909-b97ea4bef6db-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.785105 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"35c38289-b99f-44d9-9909-b97ea4bef6db","Type":"ContainerDied","Data":"a19fdea86454f7fcd5a0f759f546703e320e8bffd78ab48cd8ea8e33db695987"} Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.785726 4837 scope.go:117] "RemoveContainer" containerID="b26dda6733cd762ee75c30e5a82684efa12d85fe6e0e2dd6bc717c10ae24bf82" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.785859 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.853134 4837 scope.go:117] "RemoveContainer" containerID="51edb1e34a34ad8c44981bb1d0b3d1a436dc9b4e915aa3af6c8970e82cc36abb" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.857499 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.898289 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919135 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:30 crc kubenswrapper[4837]: E1001 09:15:30.919641 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-central-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919658 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-central-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: E1001 09:15:30.919671 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="proxy-httpd" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919678 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="proxy-httpd" Oct 01 09:15:30 crc kubenswrapper[4837]: E1001 09:15:30.919713 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-notification-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919720 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-notification-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: E1001 09:15:30.919746 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="sg-core" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919754 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="sg-core" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919936 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-notification-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919954 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="sg-core" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919968 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="ceilometer-central-agent" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.919984 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" containerName="proxy-httpd" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.921896 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.929868 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.931193 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.935019 4837 scope.go:117] "RemoveContainer" containerID="95fb7c5dfa6b87f14162660e1aa6f9d065c932873d8ee2aba70eda3c3370a785" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.937804 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.959586 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:30 crc kubenswrapper[4837]: I1001 09:15:30.967235 4837 scope.go:117] "RemoveContainer" containerID="01fb19339134679dca0f62bfab294b7a1733ecf53e1770705e42a754610aed52" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.087172 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.087363 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmstm\" (UniqueName: \"kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.087521 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.087750 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.087929 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.088044 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.088166 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.088216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190315 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmstm\" (UniqueName: \"kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190366 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190407 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190442 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190470 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190503 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190525 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.190566 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.191440 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.192538 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.195273 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.197256 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.209987 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.210253 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.210483 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmstm\" (UniqueName: \"kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.210650 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.248999 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.432251 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.796272 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerStarted","Data":"584b256413d099bc484ef115065bad072f46a82b7b9633f0218f67faf4d2e53a"} Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.826987 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35c38289-b99f-44d9-9909-b97ea4bef6db" path="/var/lib/kubelet/pods/35c38289-b99f-44d9-9909-b97ea4bef6db/volumes" Oct 01 09:15:31 crc kubenswrapper[4837]: I1001 09:15:31.862270 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.813707 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerStarted","Data":"43a533c97781838f30f31464ecaa2bc7ea99293b8e338de45539c8a6798ccd9b"} Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.813849 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-api" containerID="cri-o://a8462774cf7a801136982b3a4b5c916ba1c45069d9a6874c59d2fa526584b18c" gracePeriod=30 Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.813888 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-listener" containerID="cri-o://43a533c97781838f30f31464ecaa2bc7ea99293b8e338de45539c8a6798ccd9b" gracePeriod=30 Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.813910 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-evaluator" containerID="cri-o://91582e7ecb939bb8896ec2ec40a8583209bc646b3e57287e25b9c79db07672e9" gracePeriod=30 Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.813908 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-notifier" containerID="cri-o://584b256413d099bc484ef115065bad072f46a82b7b9633f0218f67faf4d2e53a" gracePeriod=30 Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.821101 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerStarted","Data":"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96"} Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.821153 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerStarted","Data":"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71"} Oct 01 09:15:32 crc kubenswrapper[4837]: I1001 09:15:32.821164 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerStarted","Data":"4a6125940a03cc2482ad2c2cd564433be1c6bed4f660b6e89e4c28e5a87022f7"} Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838424 4837 generic.go:334] "Generic (PLEG): container finished" podID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerID="584b256413d099bc484ef115065bad072f46a82b7b9633f0218f67faf4d2e53a" exitCode=0 Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838921 4837 generic.go:334] "Generic (PLEG): container finished" podID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerID="91582e7ecb939bb8896ec2ec40a8583209bc646b3e57287e25b9c79db07672e9" exitCode=0 Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838931 4837 generic.go:334] "Generic (PLEG): container finished" podID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerID="a8462774cf7a801136982b3a4b5c916ba1c45069d9a6874c59d2fa526584b18c" exitCode=0 Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838868 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerDied","Data":"584b256413d099bc484ef115065bad072f46a82b7b9633f0218f67faf4d2e53a"} Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838981 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerDied","Data":"91582e7ecb939bb8896ec2ec40a8583209bc646b3e57287e25b9c79db07672e9"} Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.838994 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerDied","Data":"a8462774cf7a801136982b3a4b5c916ba1c45069d9a6874c59d2fa526584b18c"} Oct 01 09:15:33 crc kubenswrapper[4837]: I1001 09:15:33.844082 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.044014228 podStartE2EDuration="7.844065785s" podCreationTimestamp="2025-10-01 09:15:26 +0000 UTC" firstStartedPulling="2025-10-01 09:15:27.271907561 +0000 UTC m=+7784.113515016" lastFinishedPulling="2025-10-01 09:15:32.071959118 +0000 UTC m=+7788.913566573" observedRunningTime="2025-10-01 09:15:32.838110411 +0000 UTC m=+7789.679717866" watchObservedRunningTime="2025-10-01 09:15:33.844065785 +0000 UTC m=+7790.685673240" Oct 01 09:15:34 crc kubenswrapper[4837]: I1001 09:15:34.851074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerStarted","Data":"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c"} Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868103 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerStarted","Data":"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930"} Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868461 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868382 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-central-agent" containerID="cri-o://cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71" gracePeriod=30 Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868460 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="proxy-httpd" containerID="cri-o://8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930" gracePeriod=30 Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868518 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-notification-agent" containerID="cri-o://4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96" gracePeriod=30 Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.868612 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="sg-core" containerID="cri-o://483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c" gracePeriod=30 Oct 01 09:15:35 crc kubenswrapper[4837]: I1001 09:15:35.914658 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.838388808 podStartE2EDuration="5.914633916s" podCreationTimestamp="2025-10-01 09:15:30 +0000 UTC" firstStartedPulling="2025-10-01 09:15:32.017246933 +0000 UTC m=+7788.858854388" lastFinishedPulling="2025-10-01 09:15:35.093492031 +0000 UTC m=+7791.935099496" observedRunningTime="2025-10-01 09:15:35.898825225 +0000 UTC m=+7792.740432710" watchObservedRunningTime="2025-10-01 09:15:35.914633916 +0000 UTC m=+7792.756241401" Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890258 4837 generic.go:334] "Generic (PLEG): container finished" podID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerID="8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930" exitCode=0 Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890300 4837 generic.go:334] "Generic (PLEG): container finished" podID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerID="483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c" exitCode=2 Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890315 4837 generic.go:334] "Generic (PLEG): container finished" podID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerID="4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96" exitCode=0 Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerDied","Data":"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930"} Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890373 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerDied","Data":"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c"} Oct 01 09:15:36 crc kubenswrapper[4837]: I1001 09:15:36.890387 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerDied","Data":"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96"} Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.489649 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660660 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660789 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmstm\" (UniqueName: \"kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660823 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660859 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660881 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.660978 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.661304 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.661354 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.661466 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.661837 4837 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.661936 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.667092 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts" (OuterVolumeSpecName: "scripts") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.686438 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm" (OuterVolumeSpecName: "kube-api-access-pmstm") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "kube-api-access-pmstm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.691809 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.734143 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.763242 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.763337 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") pod \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\" (UID: \"66dc27dd-159e-44b9-804b-4e352cb2e3eb\") " Oct 01 09:15:38 crc kubenswrapper[4837]: W1001 09:15:38.763827 4837 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/66dc27dd-159e-44b9-804b-4e352cb2e3eb/volumes/kubernetes.io~secret/combined-ca-bundle Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.763842 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764430 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764458 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764494 4837 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66dc27dd-159e-44b9-804b-4e352cb2e3eb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764508 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmstm\" (UniqueName: \"kubernetes.io/projected/66dc27dd-159e-44b9-804b-4e352cb2e3eb-kube-api-access-pmstm\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764521 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.764531 4837 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.808834 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data" (OuterVolumeSpecName: "config-data") pod "66dc27dd-159e-44b9-804b-4e352cb2e3eb" (UID: "66dc27dd-159e-44b9-804b-4e352cb2e3eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.866434 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66dc27dd-159e-44b9-804b-4e352cb2e3eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.917285 4837 generic.go:334] "Generic (PLEG): container finished" podID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerID="cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71" exitCode=0 Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.917335 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerDied","Data":"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71"} Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.917397 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66dc27dd-159e-44b9-804b-4e352cb2e3eb","Type":"ContainerDied","Data":"4a6125940a03cc2482ad2c2cd564433be1c6bed4f660b6e89e4c28e5a87022f7"} Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.917405 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.917419 4837 scope.go:117] "RemoveContainer" containerID="8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.939784 4837 scope.go:117] "RemoveContainer" containerID="483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.978781 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.987768 4837 scope.go:117] "RemoveContainer" containerID="4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96" Oct 01 09:15:38 crc kubenswrapper[4837]: I1001 09:15:38.989730 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.014635 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.015331 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-central-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015361 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-central-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.015392 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-notification-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015405 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-notification-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.015426 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="sg-core" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015437 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="sg-core" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.015471 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="proxy-httpd" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015484 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="proxy-httpd" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015821 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-notification-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015864 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="sg-core" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015887 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="proxy-httpd" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.015912 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" containerName="ceilometer-central-agent" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.019322 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.021463 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.021618 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.024361 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.027113 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.030166 4837 scope.go:117] "RemoveContainer" containerID="cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.090134 4837 scope.go:117] "RemoveContainer" containerID="8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.090490 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930\": container with ID starting with 8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930 not found: ID does not exist" containerID="8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.090536 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930"} err="failed to get container status \"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930\": rpc error: code = NotFound desc = could not find container \"8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930\": container with ID starting with 8c473614a1601627921815f4f4b10ada1b61dada0031c23d64f519fd96306930 not found: ID does not exist" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.090561 4837 scope.go:117] "RemoveContainer" containerID="483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.090892 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c\": container with ID starting with 483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c not found: ID does not exist" containerID="483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.090920 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c"} err="failed to get container status \"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c\": rpc error: code = NotFound desc = could not find container \"483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c\": container with ID starting with 483f6ecbefa1ee598283a89967c269880eee10fe52005ab8532327d55efe2f0c not found: ID does not exist" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.090970 4837 scope.go:117] "RemoveContainer" containerID="4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.091258 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96\": container with ID starting with 4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96 not found: ID does not exist" containerID="4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.091287 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96"} err="failed to get container status \"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96\": rpc error: code = NotFound desc = could not find container \"4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96\": container with ID starting with 4aa019984f5c59eda57c1a47a5368be4d444517733e0e7587200553b7f226f96 not found: ID does not exist" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.091305 4837 scope.go:117] "RemoveContainer" containerID="cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71" Oct 01 09:15:39 crc kubenswrapper[4837]: E1001 09:15:39.091572 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71\": container with ID starting with cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71 not found: ID does not exist" containerID="cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.091593 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71"} err="failed to get container status \"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71\": rpc error: code = NotFound desc = could not find container \"cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71\": container with ID starting with cf2a1ce025393efbd86d778154ecbad43bc01c8bce9455e7c040d264e8239b71 not found: ID does not exist" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.172492 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-scripts\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.172972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-run-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173039 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-config-data\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173063 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-log-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173118 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173184 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l248v\" (UniqueName: \"kubernetes.io/projected/0a41770e-165b-4d5c-9e69-8adf636ea42c-kube-api-access-l248v\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.173200 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.274660 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-scripts\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.274821 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-run-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.274911 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-config-data\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.274950 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-log-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275043 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275090 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275179 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275223 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l248v\" (UniqueName: \"kubernetes.io/projected/0a41770e-165b-4d5c-9e69-8adf636ea42c-kube-api-access-l248v\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275472 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-log-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.275519 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a41770e-165b-4d5c-9e69-8adf636ea42c-run-httpd\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.278411 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-scripts\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.279251 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-config-data\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.279464 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.279902 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.280617 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41770e-165b-4d5c-9e69-8adf636ea42c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.292999 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l248v\" (UniqueName: \"kubernetes.io/projected/0a41770e-165b-4d5c-9e69-8adf636ea42c-kube-api-access-l248v\") pod \"ceilometer-0\" (UID: \"0a41770e-165b-4d5c-9e69-8adf636ea42c\") " pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.384890 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.829300 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66dc27dd-159e-44b9-804b-4e352cb2e3eb" path="/var/lib/kubelet/pods/66dc27dd-159e-44b9-804b-4e352cb2e3eb/volumes" Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.867257 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:15:39 crc kubenswrapper[4837]: W1001 09:15:39.872924 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a41770e_165b_4d5c_9e69_8adf636ea42c.slice/crio-6d2e8656638ae63b448e7c80ca6326d27e94a02030a11bd9a6dd48cdd34c3892 WatchSource:0}: Error finding container 6d2e8656638ae63b448e7c80ca6326d27e94a02030a11bd9a6dd48cdd34c3892: Status 404 returned error can't find the container with id 6d2e8656638ae63b448e7c80ca6326d27e94a02030a11bd9a6dd48cdd34c3892 Oct 01 09:15:39 crc kubenswrapper[4837]: I1001 09:15:39.928222 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a41770e-165b-4d5c-9e69-8adf636ea42c","Type":"ContainerStarted","Data":"6d2e8656638ae63b448e7c80ca6326d27e94a02030a11bd9a6dd48cdd34c3892"} Oct 01 09:15:40 crc kubenswrapper[4837]: I1001 09:15:40.050276 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6vtpd"] Oct 01 09:15:40 crc kubenswrapper[4837]: I1001 09:15:40.062804 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6vtpd"] Oct 01 09:15:40 crc kubenswrapper[4837]: I1001 09:15:40.939965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a41770e-165b-4d5c-9e69-8adf636ea42c","Type":"ContainerStarted","Data":"ce14622a44a29549a144e63fcf7cb6d7b79837deefd3ed0daa54db0184afd451"} Oct 01 09:15:40 crc kubenswrapper[4837]: I1001 09:15:40.940282 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a41770e-165b-4d5c-9e69-8adf636ea42c","Type":"ContainerStarted","Data":"a6d812d3961d8c680c6f9be68786fd8a0673d54d7bd2a942b81df14afdb1fefc"} Oct 01 09:15:41 crc kubenswrapper[4837]: I1001 09:15:41.816465 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:15:41 crc kubenswrapper[4837]: E1001 09:15:41.816971 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:15:41 crc kubenswrapper[4837]: I1001 09:15:41.833373 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9823550a-12f9-4827-a8cc-8edb88305a2e" path="/var/lib/kubelet/pods/9823550a-12f9-4827-a8cc-8edb88305a2e/volumes" Oct 01 09:15:41 crc kubenswrapper[4837]: I1001 09:15:41.951203 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a41770e-165b-4d5c-9e69-8adf636ea42c","Type":"ContainerStarted","Data":"9188bb74a678c1f944f608409e13661b99e6cdf2e527e22c0ed8cac573605ded"} Oct 01 09:15:43 crc kubenswrapper[4837]: I1001 09:15:43.980330 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a41770e-165b-4d5c-9e69-8adf636ea42c","Type":"ContainerStarted","Data":"fc61671c3ee50dfb5475fbf48821b6ad31c4e8921bcb748da648b2aafadf31af"} Oct 01 09:15:43 crc kubenswrapper[4837]: I1001 09:15:43.980814 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:15:44 crc kubenswrapper[4837]: I1001 09:15:44.005796 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.111735822 podStartE2EDuration="6.005759486s" podCreationTimestamp="2025-10-01 09:15:38 +0000 UTC" firstStartedPulling="2025-10-01 09:15:39.875498772 +0000 UTC m=+7796.717106227" lastFinishedPulling="2025-10-01 09:15:42.769522416 +0000 UTC m=+7799.611129891" observedRunningTime="2025-10-01 09:15:43.998852594 +0000 UTC m=+7800.840460049" watchObservedRunningTime="2025-10-01 09:15:44.005759486 +0000 UTC m=+7800.847366941" Oct 01 09:15:49 crc kubenswrapper[4837]: I1001 09:15:49.041373 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-7f7c-account-create-w8s49"] Oct 01 09:15:49 crc kubenswrapper[4837]: I1001 09:15:49.058060 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-7f7c-account-create-w8s49"] Oct 01 09:15:49 crc kubenswrapper[4837]: I1001 09:15:49.837049 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ca515d7-b285-42a1-9759-6e68115bffbf" path="/var/lib/kubelet/pods/9ca515d7-b285-42a1-9759-6e68115bffbf/volumes" Oct 01 09:15:53 crc kubenswrapper[4837]: I1001 09:15:53.824779 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:15:53 crc kubenswrapper[4837]: E1001 09:15:53.825552 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.285424 4837 generic.go:334] "Generic (PLEG): container finished" podID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerID="43a533c97781838f30f31464ecaa2bc7ea99293b8e338de45539c8a6798ccd9b" exitCode=137 Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.286051 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerDied","Data":"43a533c97781838f30f31464ecaa2bc7ea99293b8e338de45539c8a6798ccd9b"} Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.286113 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"37190d42-7612-4adc-ae7f-3ce3917b3e16","Type":"ContainerDied","Data":"03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996"} Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.286128 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03354822fba1dc4e45b10f88d2ad2e59956ce83af34e14bfa6ff70815a61d996" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.324817 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.422182 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts\") pod \"37190d42-7612-4adc-ae7f-3ce3917b3e16\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.422459 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle\") pod \"37190d42-7612-4adc-ae7f-3ce3917b3e16\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.422810 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggh8t\" (UniqueName: \"kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t\") pod \"37190d42-7612-4adc-ae7f-3ce3917b3e16\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.422865 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data\") pod \"37190d42-7612-4adc-ae7f-3ce3917b3e16\" (UID: \"37190d42-7612-4adc-ae7f-3ce3917b3e16\") " Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.430826 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts" (OuterVolumeSpecName: "scripts") pod "37190d42-7612-4adc-ae7f-3ce3917b3e16" (UID: "37190d42-7612-4adc-ae7f-3ce3917b3e16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.448861 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t" (OuterVolumeSpecName: "kube-api-access-ggh8t") pod "37190d42-7612-4adc-ae7f-3ce3917b3e16" (UID: "37190d42-7612-4adc-ae7f-3ce3917b3e16"). InnerVolumeSpecName "kube-api-access-ggh8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.525571 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.525599 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggh8t\" (UniqueName: \"kubernetes.io/projected/37190d42-7612-4adc-ae7f-3ce3917b3e16-kube-api-access-ggh8t\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.579955 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data" (OuterVolumeSpecName: "config-data") pod "37190d42-7612-4adc-ae7f-3ce3917b3e16" (UID: "37190d42-7612-4adc-ae7f-3ce3917b3e16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.593076 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37190d42-7612-4adc-ae7f-3ce3917b3e16" (UID: "37190d42-7612-4adc-ae7f-3ce3917b3e16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.627237 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:03 crc kubenswrapper[4837]: I1001 09:16:03.627267 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37190d42-7612-4adc-ae7f-3ce3917b3e16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.293220 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.353349 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.362345 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.427636 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 01 09:16:04 crc kubenswrapper[4837]: E1001 09:16:04.428000 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-evaluator" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428016 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-evaluator" Oct 01 09:16:04 crc kubenswrapper[4837]: E1001 09:16:04.428039 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-api" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428046 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-api" Oct 01 09:16:04 crc kubenswrapper[4837]: E1001 09:16:04.428063 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-notifier" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428069 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-notifier" Oct 01 09:16:04 crc kubenswrapper[4837]: E1001 09:16:04.428097 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-listener" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428104 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-listener" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428268 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-listener" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428284 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-api" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428293 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-notifier" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.428310 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" containerName="aodh-evaluator" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.429995 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.435259 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-x82kr" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.436214 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.436541 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.436584 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.437407 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.473880 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543075 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-internal-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543130 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-scripts\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543174 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-config-data\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543206 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-public-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543239 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.543260 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rstvs\" (UniqueName: \"kubernetes.io/projected/4a9fe3dd-df71-4fad-9ea3-ee024052634a-kube-api-access-rstvs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644537 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-public-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644593 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644611 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rstvs\" (UniqueName: \"kubernetes.io/projected/4a9fe3dd-df71-4fad-9ea3-ee024052634a-kube-api-access-rstvs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644765 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-internal-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-scripts\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.644832 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-config-data\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.648465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-public-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.648910 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-config-data\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.649292 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-internal-tls-certs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.656906 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.657587 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9fe3dd-df71-4fad-9ea3-ee024052634a-scripts\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.668191 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rstvs\" (UniqueName: \"kubernetes.io/projected/4a9fe3dd-df71-4fad-9ea3-ee024052634a-kube-api-access-rstvs\") pod \"aodh-0\" (UID: \"4a9fe3dd-df71-4fad-9ea3-ee024052634a\") " pod="openstack/aodh-0" Oct 01 09:16:04 crc kubenswrapper[4837]: I1001 09:16:04.746965 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 01 09:16:05 crc kubenswrapper[4837]: I1001 09:16:05.217203 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 01 09:16:05 crc kubenswrapper[4837]: I1001 09:16:05.313064 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4a9fe3dd-df71-4fad-9ea3-ee024052634a","Type":"ContainerStarted","Data":"7123f5d13326c523588419e66b891379af3f1b8dc7dfc99260632df255bf1fd1"} Oct 01 09:16:05 crc kubenswrapper[4837]: I1001 09:16:05.816918 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:16:05 crc kubenswrapper[4837]: E1001 09:16:05.817493 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:05 crc kubenswrapper[4837]: I1001 09:16:05.828252 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37190d42-7612-4adc-ae7f-3ce3917b3e16" path="/var/lib/kubelet/pods/37190d42-7612-4adc-ae7f-3ce3917b3e16/volumes" Oct 01 09:16:06 crc kubenswrapper[4837]: I1001 09:16:06.327858 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4a9fe3dd-df71-4fad-9ea3-ee024052634a","Type":"ContainerStarted","Data":"501dc71ac7e6dd5d6742c20148cc0ed8520a27ba4851adae1ada8328a8e4eabd"} Oct 01 09:16:06 crc kubenswrapper[4837]: I1001 09:16:06.327898 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4a9fe3dd-df71-4fad-9ea3-ee024052634a","Type":"ContainerStarted","Data":"b8a25bb5a1fb19a673e6479a2fb6661fd1e3c2666848aa7d01841ce6a62601dd"} Oct 01 09:16:07 crc kubenswrapper[4837]: I1001 09:16:07.341357 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4a9fe3dd-df71-4fad-9ea3-ee024052634a","Type":"ContainerStarted","Data":"6d7c7dcb03ce874e4eea40614aeaa7399582031a0c5a7408891177aa1a523e48"} Oct 01 09:16:07 crc kubenswrapper[4837]: I1001 09:16:07.341647 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4a9fe3dd-df71-4fad-9ea3-ee024052634a","Type":"ContainerStarted","Data":"85f2374a730f758508e771b9b9daa0567bcd093bbfda8a537c37eb8cd08d58a7"} Oct 01 09:16:07 crc kubenswrapper[4837]: I1001 09:16:07.381431 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.9497254069999999 podStartE2EDuration="3.38141275s" podCreationTimestamp="2025-10-01 09:16:04 +0000 UTC" firstStartedPulling="2025-10-01 09:16:05.234390724 +0000 UTC m=+7822.075998199" lastFinishedPulling="2025-10-01 09:16:06.666078087 +0000 UTC m=+7823.507685542" observedRunningTime="2025-10-01 09:16:07.38019415 +0000 UTC m=+7824.221801605" watchObservedRunningTime="2025-10-01 09:16:07.38141275 +0000 UTC m=+7824.223020205" Oct 01 09:16:09 crc kubenswrapper[4837]: I1001 09:16:09.404823 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 09:16:12 crc kubenswrapper[4837]: I1001 09:16:12.280534 4837 scope.go:117] "RemoveContainer" containerID="baf2e4bc5600fceeee3a4975a44a7d65f3ce1145336e48b285792263e91ccdd8" Oct 01 09:16:12 crc kubenswrapper[4837]: I1001 09:16:12.327197 4837 scope.go:117] "RemoveContainer" containerID="beaf1bb636682fcc43dfa32c437449a18f700ec677d5e2f1163a677fe7031673" Oct 01 09:16:12 crc kubenswrapper[4837]: I1001 09:16:12.366059 4837 scope.go:117] "RemoveContainer" containerID="be001c6b3805e369ac05561f5c7aca7952a4dd023f0db6bcc76af9aa898e549f" Oct 01 09:16:12 crc kubenswrapper[4837]: I1001 09:16:12.429082 4837 scope.go:117] "RemoveContainer" containerID="9d4fd1a9c32665c0d661aac66a9c2d2778490414ea635f9e03b418d0cf33c8f9" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.237455 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784bc884c9-mzwzj"] Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.242437 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.248214 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784bc884c9-mzwzj"] Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.250532 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.338656 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784bc884c9-mzwzj"] Oct 01 09:16:14 crc kubenswrapper[4837]: E1001 09:16:14.339562 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-vtqdh openstack-cell1 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[config dns-svc kube-api-access-vtqdh openstack-cell1 ovsdbserver-nb ovsdbserver-sb]: context canceled" pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" podUID="55fd8308-bf67-4c57-9833-9b520f68aef3" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343404 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343508 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343637 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343673 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtqdh\" (UniqueName: \"kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.343778 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.361975 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.363604 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.376592 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-networker" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.384559 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.434303 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.442653 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445155 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445202 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445364 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445456 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445498 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445783 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445830 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445881 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445938 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wsk7\" (UniqueName: \"kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.445945 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446092 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446132 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446239 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446294 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtqdh\" (UniqueName: \"kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446475 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.446824 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.447083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.447554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.465614 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtqdh\" (UniqueName: \"kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh\") pod \"dnsmasq-dns-784bc884c9-mzwzj\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547502 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547705 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtqdh\" (UniqueName: \"kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547738 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547758 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547869 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.547976 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config\") pod \"55fd8308-bf67-4c57-9833-9b520f68aef3\" (UID: \"55fd8308-bf67-4c57-9833-9b520f68aef3\") " Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548002 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548290 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548332 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wsk7\" (UniqueName: \"kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548381 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548459 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548498 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548490 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548527 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548549 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config" (OuterVolumeSpecName: "config") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548620 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548632 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548644 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.548639 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.549285 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.549638 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.549661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.549638 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.550145 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.550356 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.550381 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh" (OuterVolumeSpecName: "kube-api-access-vtqdh") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "kube-api-access-vtqdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.550524 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "55fd8308-bf67-4c57-9833-9b520f68aef3" (UID: "55fd8308-bf67-4c57-9833-9b520f68aef3"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.567369 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wsk7\" (UniqueName: \"kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7\") pod \"dnsmasq-dns-f8f6dccb9-5fblz\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.649838 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.649869 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtqdh\" (UniqueName: \"kubernetes.io/projected/55fd8308-bf67-4c57-9833-9b520f68aef3-kube-api-access-vtqdh\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.649880 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/55fd8308-bf67-4c57-9833-9b520f68aef3-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:14 crc kubenswrapper[4837]: I1001 09:16:14.679451 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.055144 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-f2wtt"] Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.064429 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-f2wtt"] Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.176900 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.445968 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784bc884c9-mzwzj" Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.446047 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerStarted","Data":"c6ed6a378b7c90dcbfa40c37dbabdd57419a6c8b934a86e8d51bcb66959ed2f9"} Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.446266 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerStarted","Data":"4ecff7a70ddeb0395ee6828cf2ba1febf25e17851f59a443280d30048a25b234"} Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.695676 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784bc884c9-mzwzj"] Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.720610 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784bc884c9-mzwzj"] Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.828450 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf0317f-dd4d-40e8-aa61-5e22923f51f0" path="/var/lib/kubelet/pods/0bf0317f-dd4d-40e8-aa61-5e22923f51f0/volumes" Oct 01 09:16:15 crc kubenswrapper[4837]: I1001 09:16:15.829153 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55fd8308-bf67-4c57-9833-9b520f68aef3" path="/var/lib/kubelet/pods/55fd8308-bf67-4c57-9833-9b520f68aef3/volumes" Oct 01 09:16:16 crc kubenswrapper[4837]: I1001 09:16:16.464808 4837 generic.go:334] "Generic (PLEG): container finished" podID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerID="c6ed6a378b7c90dcbfa40c37dbabdd57419a6c8b934a86e8d51bcb66959ed2f9" exitCode=0 Oct 01 09:16:16 crc kubenswrapper[4837]: I1001 09:16:16.464890 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerDied","Data":"c6ed6a378b7c90dcbfa40c37dbabdd57419a6c8b934a86e8d51bcb66959ed2f9"} Oct 01 09:16:16 crc kubenswrapper[4837]: I1001 09:16:16.464970 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerStarted","Data":"c15da86727a72aadd1057cbeeebb45a9abf3e9ac3df1ee3eeaeff08d482e0740"} Oct 01 09:16:16 crc kubenswrapper[4837]: I1001 09:16:16.465138 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:16 crc kubenswrapper[4837]: I1001 09:16:16.504223 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" podStartSLOduration=2.504197281 podStartE2EDuration="2.504197281s" podCreationTimestamp="2025-10-01 09:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:16:16.492169952 +0000 UTC m=+7833.333777457" watchObservedRunningTime="2025-10-01 09:16:16.504197281 +0000 UTC m=+7833.345804776" Oct 01 09:16:17 crc kubenswrapper[4837]: I1001 09:16:17.816353 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:16:17 crc kubenswrapper[4837]: E1001 09:16:17.816883 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.681930 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.759825 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.760142 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="dnsmasq-dns" containerID="cri-o://07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f" gracePeriod=10 Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.959903 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.962293 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:24 crc kubenswrapper[4837]: I1001 09:16:24.974624 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108722 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108766 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108804 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108865 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6njxd\" (UniqueName: \"kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108930 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.108948 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210082 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210124 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210165 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210224 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210275 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6njxd\" (UniqueName: \"kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210293 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.210311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.211198 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.211315 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.211742 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.211915 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.212382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.212573 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.239247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6njxd\" (UniqueName: \"kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd\") pod \"dnsmasq-dns-5f96d8f66f-72bc4\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.291622 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.403082 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.516723 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.516842 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdfhp\" (UniqueName: \"kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.516917 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.516939 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.517089 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.580127 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp" (OuterVolumeSpecName: "kube-api-access-bdfhp") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "kube-api-access-bdfhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.599935 4837 generic.go:334] "Generic (PLEG): container finished" podID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerID="07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f" exitCode=0 Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.600129 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" event={"ID":"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09","Type":"ContainerDied","Data":"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f"} Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.600248 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" event={"ID":"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09","Type":"ContainerDied","Data":"467bbdbb1cc8353a2cc109d148ec841ec8df4fe7288c9c890dcbddd67acbae79"} Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.600338 4837 scope.go:117] "RemoveContainer" containerID="07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.600500 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b74c447c-79rbh" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.620344 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.624190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") pod \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\" (UID: \"8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09\") " Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.624855 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdfhp\" (UniqueName: \"kubernetes.io/projected/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-kube-api-access-bdfhp\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:25 crc kubenswrapper[4837]: W1001 09:16:25.625090 4837 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.625104 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.629798 4837 scope.go:117] "RemoveContainer" containerID="c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.633121 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.638559 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config" (OuterVolumeSpecName: "config") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.641626 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" (UID: "8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.669663 4837 scope.go:117] "RemoveContainer" containerID="07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f" Oct 01 09:16:25 crc kubenswrapper[4837]: E1001 09:16:25.672307 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f\": container with ID starting with 07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f not found: ID does not exist" containerID="07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.672348 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f"} err="failed to get container status \"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f\": rpc error: code = NotFound desc = could not find container \"07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f\": container with ID starting with 07018893fc9ed527e4873acc192286217de199989369c54bd0ffc15dbdbf4f6f not found: ID does not exist" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.672375 4837 scope.go:117] "RemoveContainer" containerID="c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6" Oct 01 09:16:25 crc kubenswrapper[4837]: E1001 09:16:25.672771 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6\": container with ID starting with c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6 not found: ID does not exist" containerID="c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.672789 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6"} err="failed to get container status \"c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6\": rpc error: code = NotFound desc = could not find container \"c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6\": container with ID starting with c77ad706713d27f52e80c9af3d2ccd0267479b1af7a8ac7dfdebcf1e1ed370a6 not found: ID does not exist" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.727686 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.728230 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.728242 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.728253 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.962277 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.970886 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:16:25 crc kubenswrapper[4837]: I1001 09:16:25.980370 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57b74c447c-79rbh"] Oct 01 09:16:26 crc kubenswrapper[4837]: I1001 09:16:26.613621 4837 generic.go:334] "Generic (PLEG): container finished" podID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerID="e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1" exitCode=0 Oct 01 09:16:26 crc kubenswrapper[4837]: I1001 09:16:26.613826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" event={"ID":"1fb91776-8eac-49e7-9dba-5ab6ffd37938","Type":"ContainerDied","Data":"e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1"} Oct 01 09:16:26 crc kubenswrapper[4837]: I1001 09:16:26.614957 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" event={"ID":"1fb91776-8eac-49e7-9dba-5ab6ffd37938","Type":"ContainerStarted","Data":"7e096cb7f2f815f794e874fb611b70748a173c920e77d12673d05a367230c9f0"} Oct 01 09:16:27 crc kubenswrapper[4837]: I1001 09:16:27.644928 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" event={"ID":"1fb91776-8eac-49e7-9dba-5ab6ffd37938","Type":"ContainerStarted","Data":"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd"} Oct 01 09:16:27 crc kubenswrapper[4837]: I1001 09:16:27.645665 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:27 crc kubenswrapper[4837]: I1001 09:16:27.674453 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" podStartSLOduration=3.674433248 podStartE2EDuration="3.674433248s" podCreationTimestamp="2025-10-01 09:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:16:27.665678341 +0000 UTC m=+7844.507285816" watchObservedRunningTime="2025-10-01 09:16:27.674433248 +0000 UTC m=+7844.516040703" Oct 01 09:16:27 crc kubenswrapper[4837]: I1001 09:16:27.838173 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" path="/var/lib/kubelet/pods/8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09/volumes" Oct 01 09:16:28 crc kubenswrapper[4837]: I1001 09:16:28.817147 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:16:28 crc kubenswrapper[4837]: E1001 09:16:28.817378 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.294301 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.384754 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.385017 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="dnsmasq-dns" containerID="cri-o://c15da86727a72aadd1057cbeeebb45a9abf3e9ac3df1ee3eeaeff08d482e0740" gracePeriod=10 Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.570239 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 09:16:35 crc kubenswrapper[4837]: E1001 09:16:35.580814 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="dnsmasq-dns" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.580848 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="dnsmasq-dns" Oct 01 09:16:35 crc kubenswrapper[4837]: E1001 09:16:35.580887 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="init" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.580896 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="init" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.581320 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa0abb8-7138-4cc0-9c27-3cbf25a6bd09" containerName="dnsmasq-dns" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.582597 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.582721 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686316 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686413 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnw2p\" (UniqueName: \"kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686470 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686765 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.686913 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.687001 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.731319 4837 generic.go:334] "Generic (PLEG): container finished" podID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerID="c15da86727a72aadd1057cbeeebb45a9abf3e9ac3df1ee3eeaeff08d482e0740" exitCode=0 Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.731373 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerDied","Data":"c15da86727a72aadd1057cbeeebb45a9abf3e9ac3df1ee3eeaeff08d482e0740"} Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789322 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnw2p\" (UniqueName: \"kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789629 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789650 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789721 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789759 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789784 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.789830 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.791542 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.792074 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.792112 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.792900 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.793060 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.793764 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.809069 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnw2p\" (UniqueName: \"kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p\") pod \"dnsmasq-dns-6d8594d5ff-7nl6w\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:35 crc kubenswrapper[4837]: I1001 09:16:35.904471 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.057631 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197164 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wsk7\" (UniqueName: \"kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197508 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197532 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197591 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197659 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197757 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.197861 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config\") pod \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\" (UID: \"0d54ff47-c0b1-4ba7-9389-d3936e05994c\") " Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.203952 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7" (OuterVolumeSpecName: "kube-api-access-8wsk7") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "kube-api-access-8wsk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.248650 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.250247 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.255717 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config" (OuterVolumeSpecName: "config") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.257389 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.260093 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.260687 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d54ff47-c0b1-4ba7-9389-d3936e05994c" (UID: "0d54ff47-c0b1-4ba7-9389-d3936e05994c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.301201 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wsk7\" (UniqueName: \"kubernetes.io/projected/0d54ff47-c0b1-4ba7-9389-d3936e05994c-kube-api-access-8wsk7\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302181 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302269 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302346 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302449 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302528 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.302612 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d54ff47-c0b1-4ba7-9389-d3936e05994c-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.421899 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.753282 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" event={"ID":"0d54ff47-c0b1-4ba7-9389-d3936e05994c","Type":"ContainerDied","Data":"4ecff7a70ddeb0395ee6828cf2ba1febf25e17851f59a443280d30048a25b234"} Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.753560 4837 scope.go:117] "RemoveContainer" containerID="c15da86727a72aadd1057cbeeebb45a9abf3e9ac3df1ee3eeaeff08d482e0740" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.753391 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f8f6dccb9-5fblz" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.758562 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" event={"ID":"2f6ede11-1c94-4644-8a8d-b508c1d48cca","Type":"ContainerStarted","Data":"e6cf9d1fe3404cd95aba6aa5ab1f36d30fe23ee79a96ee2645bc83fc9ae04cd8"} Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.777060 4837 scope.go:117] "RemoveContainer" containerID="c6ed6a378b7c90dcbfa40c37dbabdd57419a6c8b934a86e8d51bcb66959ed2f9" Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.795301 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:36 crc kubenswrapper[4837]: I1001 09:16:36.805846 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f8f6dccb9-5fblz"] Oct 01 09:16:37 crc kubenswrapper[4837]: I1001 09:16:37.774640 4837 generic.go:334] "Generic (PLEG): container finished" podID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerID="df08dd925c5831cb249c3a3bd952809b010c8c747556c264b64ae57c5481ca6d" exitCode=0 Oct 01 09:16:37 crc kubenswrapper[4837]: I1001 09:16:37.774731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" event={"ID":"2f6ede11-1c94-4644-8a8d-b508c1d48cca","Type":"ContainerDied","Data":"df08dd925c5831cb249c3a3bd952809b010c8c747556c264b64ae57c5481ca6d"} Oct 01 09:16:37 crc kubenswrapper[4837]: I1001 09:16:37.827640 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" path="/var/lib/kubelet/pods/0d54ff47-c0b1-4ba7-9389-d3936e05994c/volumes" Oct 01 09:16:38 crc kubenswrapper[4837]: I1001 09:16:38.791580 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" event={"ID":"2f6ede11-1c94-4644-8a8d-b508c1d48cca","Type":"ContainerStarted","Data":"7587a55738e24cf1357a086d0af7f30cf8e438a689be538577d1f89ebd3e96e1"} Oct 01 09:16:38 crc kubenswrapper[4837]: I1001 09:16:38.792276 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:38 crc kubenswrapper[4837]: I1001 09:16:38.840803 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" podStartSLOduration=3.84077538 podStartE2EDuration="3.84077538s" podCreationTimestamp="2025-10-01 09:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:16:38.821796309 +0000 UTC m=+7855.663403854" watchObservedRunningTime="2025-10-01 09:16:38.84077538 +0000 UTC m=+7855.682382875" Oct 01 09:16:39 crc kubenswrapper[4837]: I1001 09:16:39.816026 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:16:39 crc kubenswrapper[4837]: E1001 09:16:39.816367 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.812102 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:16:44 crc kubenswrapper[4837]: E1001 09:16:44.813226 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="init" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.813250 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="init" Oct 01 09:16:44 crc kubenswrapper[4837]: E1001 09:16:44.813309 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="dnsmasq-dns" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.813327 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="dnsmasq-dns" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.813748 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d54ff47-c0b1-4ba7-9389-d3936e05994c" containerName="dnsmasq-dns" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.816577 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.840175 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.949560 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.949678 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:44 crc kubenswrapper[4837]: I1001 09:16:44.949968 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7rks\" (UniqueName: \"kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.051608 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7rks\" (UniqueName: \"kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.051950 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.052096 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.052590 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.052639 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.084875 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7rks\" (UniqueName: \"kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks\") pod \"certified-operators-8kklb\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.162465 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.660187 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.893253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerStarted","Data":"e3966e7ac8ea06aa663e295ba69d0e6f1e83b76fb8f2751aab81849d33c2297d"} Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.905993 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.991576 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:45 crc kubenswrapper[4837]: I1001 09:16:45.991817 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="dnsmasq-dns" containerID="cri-o://0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd" gracePeriod=10 Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.419575 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh"] Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.421113 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.423797 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.424518 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.424720 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.429270 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.442793 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t"] Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.444174 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.446052 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.446519 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.453304 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh"] Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.464042 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t"] Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483731 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483778 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483835 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483869 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xmvm\" (UniqueName: \"kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483890 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483908 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.483996 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l25p2\" (UniqueName: \"kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.484013 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.579983 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585749 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l25p2\" (UniqueName: \"kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585780 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585848 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585868 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585918 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585952 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xmvm\" (UniqueName: \"kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585974 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.585993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.592120 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.593622 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.597359 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.604490 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.605583 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.607370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.618247 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xmvm\" (UniqueName: \"kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.632329 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l25p2\" (UniqueName: \"kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n2924t\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687679 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687741 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687758 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6njxd\" (UniqueName: \"kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687863 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687893 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.687964 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1\") pod \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\" (UID: \"1fb91776-8eac-49e7-9dba-5ab6ffd37938\") " Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.707926 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd" (OuterVolumeSpecName: "kube-api-access-6njxd") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "kube-api-access-6njxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.739082 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.746608 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.747640 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.752843 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.753589 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.755284 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config" (OuterVolumeSpecName: "config") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.777651 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.777873 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "1fb91776-8eac-49e7-9dba-5ab6ffd37938" (UID: "1fb91776-8eac-49e7-9dba-5ab6ffd37938"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.789990 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.790150 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.790207 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.790856 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.790931 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6njxd\" (UniqueName: \"kubernetes.io/projected/1fb91776-8eac-49e7-9dba-5ab6ffd37938-kube-api-access-6njxd\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.790985 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.791064 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fb91776-8eac-49e7-9dba-5ab6ffd37938-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.917143 4837 generic.go:334] "Generic (PLEG): container finished" podID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerID="0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd" exitCode=0 Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.917232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" event={"ID":"1fb91776-8eac-49e7-9dba-5ab6ffd37938","Type":"ContainerDied","Data":"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd"} Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.917259 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" event={"ID":"1fb91776-8eac-49e7-9dba-5ab6ffd37938","Type":"ContainerDied","Data":"7e096cb7f2f815f794e874fb611b70748a173c920e77d12673d05a367230c9f0"} Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.917276 4837 scope.go:117] "RemoveContainer" containerID="0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.917407 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f96d8f66f-72bc4" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.925358 4837 generic.go:334] "Generic (PLEG): container finished" podID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerID="8d779aec4868120c882adc08c1c873dc2853ba3fe1f7e05630a935c1d9989a7c" exitCode=0 Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.925397 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerDied","Data":"8d779aec4868120c882adc08c1c873dc2853ba3fe1f7e05630a935c1d9989a7c"} Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.945967 4837 scope.go:117] "RemoveContainer" containerID="e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.972771 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.972903 4837 scope.go:117] "RemoveContainer" containerID="0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd" Oct 01 09:16:46 crc kubenswrapper[4837]: E1001 09:16:46.977863 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd\": container with ID starting with 0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd not found: ID does not exist" containerID="0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.977907 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd"} err="failed to get container status \"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd\": rpc error: code = NotFound desc = could not find container \"0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd\": container with ID starting with 0b24b5b0d82bdecee07810746e2c79f2959ed8a96c409a8276de5ffbacc10bbd not found: ID does not exist" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.977936 4837 scope.go:117] "RemoveContainer" containerID="e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1" Oct 01 09:16:46 crc kubenswrapper[4837]: E1001 09:16:46.978277 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1\": container with ID starting with e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1 not found: ID does not exist" containerID="e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.978308 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1"} err="failed to get container status \"e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1\": rpc error: code = NotFound desc = could not find container \"e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1\": container with ID starting with e056a479efe8e74cfa1f13cec3436b759d30ab4e7e957e5e8cea662bf110eef1 not found: ID does not exist" Oct 01 09:16:46 crc kubenswrapper[4837]: I1001 09:16:46.984060 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f96d8f66f-72bc4"] Oct 01 09:16:47 crc kubenswrapper[4837]: I1001 09:16:47.642467 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t"] Oct 01 09:16:47 crc kubenswrapper[4837]: W1001 09:16:47.659110 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9f6cd02_35a4_4528_abd6_70d505abe8c3.slice/crio-302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f WatchSource:0}: Error finding container 302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f: Status 404 returned error can't find the container with id 302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f Oct 01 09:16:47 crc kubenswrapper[4837]: I1001 09:16:47.834830 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" path="/var/lib/kubelet/pods/1fb91776-8eac-49e7-9dba-5ab6ffd37938/volumes" Oct 01 09:16:47 crc kubenswrapper[4837]: I1001 09:16:47.960986 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" event={"ID":"e9f6cd02-35a4-4528-abd6-70d505abe8c3","Type":"ContainerStarted","Data":"302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f"} Oct 01 09:16:48 crc kubenswrapper[4837]: I1001 09:16:48.312023 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh"] Oct 01 09:16:48 crc kubenswrapper[4837]: I1001 09:16:48.976796 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerStarted","Data":"394f6af21f4b50354f4acbd0852186864979bd7858f7fa1f11441792548e77dd"} Oct 01 09:16:48 crc kubenswrapper[4837]: I1001 09:16:48.979486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" event={"ID":"d99bc7f6-69da-47d6-8a2b-0efeed9445cf","Type":"ContainerStarted","Data":"e14146946df6bbf486eb4722680cd547f2a27b3603cd1970662b72ae2b964517"} Oct 01 09:16:49 crc kubenswrapper[4837]: I1001 09:16:49.990786 4837 generic.go:334] "Generic (PLEG): container finished" podID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerID="394f6af21f4b50354f4acbd0852186864979bd7858f7fa1f11441792548e77dd" exitCode=0 Oct 01 09:16:49 crc kubenswrapper[4837]: I1001 09:16:49.990840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerDied","Data":"394f6af21f4b50354f4acbd0852186864979bd7858f7fa1f11441792548e77dd"} Oct 01 09:16:50 crc kubenswrapper[4837]: I1001 09:16:50.057773 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-6m6cr"] Oct 01 09:16:50 crc kubenswrapper[4837]: I1001 09:16:50.066563 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-6m6cr"] Oct 01 09:16:51 crc kubenswrapper[4837]: I1001 09:16:51.816911 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:16:51 crc kubenswrapper[4837]: E1001 09:16:51.817249 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:16:51 crc kubenswrapper[4837]: I1001 09:16:51.829542 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700833d9-e163-4dfa-b285-1493e045e28e" path="/var/lib/kubelet/pods/700833d9-e163-4dfa-b285-1493e045e28e/volumes" Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.104109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerStarted","Data":"a7f2291be7d22a76fba9a3d959ddc6ef518dcf253635ac4bbde334cd4db9fc6f"} Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.106134 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" event={"ID":"d99bc7f6-69da-47d6-8a2b-0efeed9445cf","Type":"ContainerStarted","Data":"10f3e3699644f214b302d8c7ca035aec07a82e0d2f414ff843842a4f64022ae4"} Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.109528 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" event={"ID":"e9f6cd02-35a4-4528-abd6-70d505abe8c3","Type":"ContainerStarted","Data":"f500d9dfc871fc3d03b4d050404effeafc041540b8bbafbf118d246209a72ca1"} Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.131866 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8kklb" podStartSLOduration=4.385680389 podStartE2EDuration="14.131847734s" podCreationTimestamp="2025-10-01 09:16:44 +0000 UTC" firstStartedPulling="2025-10-01 09:16:46.93196125 +0000 UTC m=+7863.773568745" lastFinishedPulling="2025-10-01 09:16:56.678128595 +0000 UTC m=+7873.519736090" observedRunningTime="2025-10-01 09:16:58.124253255 +0000 UTC m=+7874.965860720" watchObservedRunningTime="2025-10-01 09:16:58.131847734 +0000 UTC m=+7874.973455199" Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.148550 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" podStartSLOduration=3.034616445 podStartE2EDuration="12.148532196s" podCreationTimestamp="2025-10-01 09:16:46 +0000 UTC" firstStartedPulling="2025-10-01 09:16:47.66409303 +0000 UTC m=+7864.505700495" lastFinishedPulling="2025-10-01 09:16:56.778008791 +0000 UTC m=+7873.619616246" observedRunningTime="2025-10-01 09:16:58.146281591 +0000 UTC m=+7874.987889056" watchObservedRunningTime="2025-10-01 09:16:58.148532196 +0000 UTC m=+7874.990139651" Oct 01 09:16:58 crc kubenswrapper[4837]: I1001 09:16:58.172450 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" podStartSLOduration=3.7055506019999997 podStartE2EDuration="12.17243112s" podCreationTimestamp="2025-10-01 09:16:46 +0000 UTC" firstStartedPulling="2025-10-01 09:16:48.323877298 +0000 UTC m=+7865.165484773" lastFinishedPulling="2025-10-01 09:16:56.790757836 +0000 UTC m=+7873.632365291" observedRunningTime="2025-10-01 09:16:58.161707684 +0000 UTC m=+7875.003315139" watchObservedRunningTime="2025-10-01 09:16:58.17243112 +0000 UTC m=+7875.014038575" Oct 01 09:17:00 crc kubenswrapper[4837]: I1001 09:17:00.047272 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9a77-account-create-jsw66"] Oct 01 09:17:00 crc kubenswrapper[4837]: I1001 09:17:00.061336 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9a77-account-create-jsw66"] Oct 01 09:17:01 crc kubenswrapper[4837]: I1001 09:17:01.833681 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512b846a-2ff8-44e3-b22f-f27dbe4d1571" path="/var/lib/kubelet/pods/512b846a-2ff8-44e3-b22f-f27dbe4d1571/volumes" Oct 01 09:17:03 crc kubenswrapper[4837]: I1001 09:17:03.825139 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:17:03 crc kubenswrapper[4837]: E1001 09:17:03.825892 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:17:05 crc kubenswrapper[4837]: I1001 09:17:05.163151 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:05 crc kubenswrapper[4837]: I1001 09:17:05.163474 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:06 crc kubenswrapper[4837]: I1001 09:17:06.240542 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8kklb" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="registry-server" probeResult="failure" output=< Oct 01 09:17:06 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:17:06 crc kubenswrapper[4837]: > Oct 01 09:17:07 crc kubenswrapper[4837]: I1001 09:17:07.211846 4837 generic.go:334] "Generic (PLEG): container finished" podID="e9f6cd02-35a4-4528-abd6-70d505abe8c3" containerID="f500d9dfc871fc3d03b4d050404effeafc041540b8bbafbf118d246209a72ca1" exitCode=0 Oct 01 09:17:07 crc kubenswrapper[4837]: I1001 09:17:07.211945 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" event={"ID":"e9f6cd02-35a4-4528-abd6-70d505abe8c3","Type":"ContainerDied","Data":"f500d9dfc871fc3d03b4d050404effeafc041540b8bbafbf118d246209a72ca1"} Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.227169 4837 generic.go:334] "Generic (PLEG): container finished" podID="d99bc7f6-69da-47d6-8a2b-0efeed9445cf" containerID="10f3e3699644f214b302d8c7ca035aec07a82e0d2f414ff843842a4f64022ae4" exitCode=0 Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.227261 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" event={"ID":"d99bc7f6-69da-47d6-8a2b-0efeed9445cf","Type":"ContainerDied","Data":"10f3e3699644f214b302d8c7ca035aec07a82e0d2f414ff843842a4f64022ae4"} Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.770139 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.923464 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle\") pod \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.923656 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key\") pod \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.924951 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory\") pod \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.925364 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l25p2\" (UniqueName: \"kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2\") pod \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\" (UID: \"e9f6cd02-35a4-4528-abd6-70d505abe8c3\") " Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.931959 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "e9f6cd02-35a4-4528-abd6-70d505abe8c3" (UID: "e9f6cd02-35a4-4528-abd6-70d505abe8c3"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.932178 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2" (OuterVolumeSpecName: "kube-api-access-l25p2") pod "e9f6cd02-35a4-4528-abd6-70d505abe8c3" (UID: "e9f6cd02-35a4-4528-abd6-70d505abe8c3"). InnerVolumeSpecName "kube-api-access-l25p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.976561 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e9f6cd02-35a4-4528-abd6-70d505abe8c3" (UID: "e9f6cd02-35a4-4528-abd6-70d505abe8c3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:08 crc kubenswrapper[4837]: I1001 09:17:08.980860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory" (OuterVolumeSpecName: "inventory") pod "e9f6cd02-35a4-4528-abd6-70d505abe8c3" (UID: "e9f6cd02-35a4-4528-abd6-70d505abe8c3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.028963 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.029017 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l25p2\" (UniqueName: \"kubernetes.io/projected/e9f6cd02-35a4-4528-abd6-70d505abe8c3-kube-api-access-l25p2\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.029035 4837 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.029050 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9f6cd02-35a4-4528-abd6-70d505abe8c3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.251400 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" event={"ID":"e9f6cd02-35a4-4528-abd6-70d505abe8c3","Type":"ContainerDied","Data":"302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f"} Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.251468 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="302c2d5a302af224558576a8610f7811e6ff2ed1f488afab6c1791f820e0974f" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.251470 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n2924t" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.788635 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.951457 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xmvm\" (UniqueName: \"kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm\") pod \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.951527 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle\") pod \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.951612 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory\") pod \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.951952 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key\") pod \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\" (UID: \"d99bc7f6-69da-47d6-8a2b-0efeed9445cf\") " Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.957081 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm" (OuterVolumeSpecName: "kube-api-access-7xmvm") pod "d99bc7f6-69da-47d6-8a2b-0efeed9445cf" (UID: "d99bc7f6-69da-47d6-8a2b-0efeed9445cf"). InnerVolumeSpecName "kube-api-access-7xmvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.957413 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "d99bc7f6-69da-47d6-8a2b-0efeed9445cf" (UID: "d99bc7f6-69da-47d6-8a2b-0efeed9445cf"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.988480 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d99bc7f6-69da-47d6-8a2b-0efeed9445cf" (UID: "d99bc7f6-69da-47d6-8a2b-0efeed9445cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:09 crc kubenswrapper[4837]: I1001 09:17:09.996687 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory" (OuterVolumeSpecName: "inventory") pod "d99bc7f6-69da-47d6-8a2b-0efeed9445cf" (UID: "d99bc7f6-69da-47d6-8a2b-0efeed9445cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.054528 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-n277b"] Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.056844 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.056896 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xmvm\" (UniqueName: \"kubernetes.io/projected/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-kube-api-access-7xmvm\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.056917 4837 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.056938 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d99bc7f6-69da-47d6-8a2b-0efeed9445cf-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.062241 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-n277b"] Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.268965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" event={"ID":"d99bc7f6-69da-47d6-8a2b-0efeed9445cf","Type":"ContainerDied","Data":"e14146946df6bbf486eb4722680cd547f2a27b3603cd1970662b72ae2b964517"} Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.269021 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e14146946df6bbf486eb4722680cd547f2a27b3603cd1970662b72ae2b964517" Oct 01 09:17:10 crc kubenswrapper[4837]: I1001 09:17:10.269070 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh" Oct 01 09:17:11 crc kubenswrapper[4837]: I1001 09:17:11.838563 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b96552d-f2ac-43c9-9d2e-786da4133388" path="/var/lib/kubelet/pods/4b96552d-f2ac-43c9-9d2e-786da4133388/volumes" Oct 01 09:17:12 crc kubenswrapper[4837]: I1001 09:17:12.640221 4837 scope.go:117] "RemoveContainer" containerID="ca4aed4b33ab5da98a7b94d1d093d97e7bfd7448ca44a7ba5ed34030ec3f3af8" Oct 01 09:17:12 crc kubenswrapper[4837]: I1001 09:17:12.708252 4837 scope.go:117] "RemoveContainer" containerID="3341089752c8c88877dd6c27557df0e2b1fffeca21428dfe799f64441bbbaffe" Oct 01 09:17:12 crc kubenswrapper[4837]: I1001 09:17:12.934841 4837 scope.go:117] "RemoveContainer" containerID="f7b842d57e503b7e204b4b9518b076623531d2074b4518c54b8e81269b8d4112" Oct 01 09:17:12 crc kubenswrapper[4837]: I1001 09:17:12.962459 4837 scope.go:117] "RemoveContainer" containerID="6aa3e91be97793fe408de8982b33513f0bf73fa474a058382879348a8a4a77c5" Oct 01 09:17:13 crc kubenswrapper[4837]: I1001 09:17:13.037784 4837 scope.go:117] "RemoveContainer" containerID="9c04d6d846b070c5bfba0beccce8fa1b061aaf63d55e43edf65d61e957f58c90" Oct 01 09:17:13 crc kubenswrapper[4837]: I1001 09:17:13.068472 4837 scope.go:117] "RemoveContainer" containerID="722355f307a5bd3e6b2812b6a2eb07fa4571e3767b5a480aec16291c47a68e43" Oct 01 09:17:13 crc kubenswrapper[4837]: I1001 09:17:13.155362 4837 scope.go:117] "RemoveContainer" containerID="99fdcae14a1692b6ffa7e89638b0b14c7f8c0504163cdf0d012c71fb5f50a9f2" Oct 01 09:17:13 crc kubenswrapper[4837]: I1001 09:17:13.202750 4837 scope.go:117] "RemoveContainer" containerID="a0d1e139373f3331c9de3b2a407fe939f1a63d2bf4690bc47f859170915dd769" Oct 01 09:17:15 crc kubenswrapper[4837]: I1001 09:17:15.246637 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:15 crc kubenswrapper[4837]: I1001 09:17:15.325826 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:15 crc kubenswrapper[4837]: I1001 09:17:15.815866 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:17:15 crc kubenswrapper[4837]: E1001 09:17:15.816272 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:17:16 crc kubenswrapper[4837]: I1001 09:17:16.022171 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:17:16 crc kubenswrapper[4837]: I1001 09:17:16.347929 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8kklb" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="registry-server" containerID="cri-o://a7f2291be7d22a76fba9a3d959ddc6ef518dcf253635ac4bbde334cd4db9fc6f" gracePeriod=2 Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.361471 4837 generic.go:334] "Generic (PLEG): container finished" podID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerID="a7f2291be7d22a76fba9a3d959ddc6ef518dcf253635ac4bbde334cd4db9fc6f" exitCode=0 Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.361546 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerDied","Data":"a7f2291be7d22a76fba9a3d959ddc6ef518dcf253635ac4bbde334cd4db9fc6f"} Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.361958 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kklb" event={"ID":"aef7bac5-3ec7-4348-8401-d7f8ab52be4c","Type":"ContainerDied","Data":"e3966e7ac8ea06aa663e295ba69d0e6f1e83b76fb8f2751aab81849d33c2297d"} Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.361983 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3966e7ac8ea06aa663e295ba69d0e6f1e83b76fb8f2751aab81849d33c2297d" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.424737 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.467795 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content\") pod \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.467895 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7rks\" (UniqueName: \"kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks\") pod \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.468029 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities\") pod \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\" (UID: \"aef7bac5-3ec7-4348-8401-d7f8ab52be4c\") " Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.469471 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities" (OuterVolumeSpecName: "utilities") pod "aef7bac5-3ec7-4348-8401-d7f8ab52be4c" (UID: "aef7bac5-3ec7-4348-8401-d7f8ab52be4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.475969 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks" (OuterVolumeSpecName: "kube-api-access-r7rks") pod "aef7bac5-3ec7-4348-8401-d7f8ab52be4c" (UID: "aef7bac5-3ec7-4348-8401-d7f8ab52be4c"). InnerVolumeSpecName "kube-api-access-r7rks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.534184 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aef7bac5-3ec7-4348-8401-d7f8ab52be4c" (UID: "aef7bac5-3ec7-4348-8401-d7f8ab52be4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.570832 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.570868 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7rks\" (UniqueName: \"kubernetes.io/projected/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-kube-api-access-r7rks\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:17 crc kubenswrapper[4837]: I1001 09:17:17.570883 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aef7bac5-3ec7-4348-8401-d7f8ab52be4c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:17:18 crc kubenswrapper[4837]: I1001 09:17:18.384538 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kklb" Oct 01 09:17:18 crc kubenswrapper[4837]: I1001 09:17:18.432886 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:17:18 crc kubenswrapper[4837]: I1001 09:17:18.447196 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8kklb"] Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.249688 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm"] Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250207 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="dnsmasq-dns" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250233 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="dnsmasq-dns" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250249 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99bc7f6-69da-47d6-8a2b-0efeed9445cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250263 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99bc7f6-69da-47d6-8a2b-0efeed9445cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250272 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="init" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250279 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="init" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250299 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="registry-server" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250306 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="registry-server" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250322 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9f6cd02-35a4-4528-abd6-70d505abe8c3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250332 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9f6cd02-35a4-4528-abd6-70d505abe8c3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250358 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="extract-utilities" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250366 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="extract-utilities" Oct 01 09:17:19 crc kubenswrapper[4837]: E1001 09:17:19.250383 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="extract-content" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.250390 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="extract-content" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.251353 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" containerName="registry-server" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.251409 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d99bc7f6-69da-47d6-8a2b-0efeed9445cf" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.251426 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb91776-8eac-49e7-9dba-5ab6ffd37938" containerName="dnsmasq-dns" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.251437 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9f6cd02-35a4-4528-abd6-70d505abe8c3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.252330 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.255732 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.256712 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.263445 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.264109 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.268923 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t"] Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.271176 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.273933 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.276540 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.283756 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm"] Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.303107 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t"] Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314215 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314324 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314434 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btmhg\" (UniqueName: \"kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314666 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314739 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pwn8\" (UniqueName: \"kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314794 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.314902 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417028 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417108 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417154 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417189 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btmhg\" (UniqueName: \"kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417276 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417304 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.417329 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pwn8\" (UniqueName: \"kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.425422 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.427170 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.428570 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.429340 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.430131 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.430590 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.436128 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btmhg\" (UniqueName: \"kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.449846 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pwn8\" (UniqueName: \"kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.575729 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.589928 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:17:19 crc kubenswrapper[4837]: I1001 09:17:19.831915 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef7bac5-3ec7-4348-8401-d7f8ab52be4c" path="/var/lib/kubelet/pods/aef7bac5-3ec7-4348-8401-d7f8ab52be4c/volumes" Oct 01 09:17:20 crc kubenswrapper[4837]: I1001 09:17:20.185735 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t"] Oct 01 09:17:20 crc kubenswrapper[4837]: W1001 09:17:20.190299 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e10153b_fcd7_432f_95c2_c240f6b53944.slice/crio-a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa WatchSource:0}: Error finding container a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa: Status 404 returned error can't find the container with id a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa Oct 01 09:17:20 crc kubenswrapper[4837]: W1001 09:17:20.280979 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc52d12_9bf2_4649_9ea5_68d196c129a4.slice/crio-de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508 WatchSource:0}: Error finding container de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508: Status 404 returned error can't find the container with id de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508 Oct 01 09:17:20 crc kubenswrapper[4837]: I1001 09:17:20.281115 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm"] Oct 01 09:17:20 crc kubenswrapper[4837]: I1001 09:17:20.403766 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" event={"ID":"3e10153b-fcd7-432f-95c2-c240f6b53944","Type":"ContainerStarted","Data":"a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa"} Oct 01 09:17:20 crc kubenswrapper[4837]: I1001 09:17:20.405316 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" event={"ID":"5fc52d12-9bf2-4649-9ea5-68d196c129a4","Type":"ContainerStarted","Data":"de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508"} Oct 01 09:17:21 crc kubenswrapper[4837]: I1001 09:17:21.421726 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" event={"ID":"5fc52d12-9bf2-4649-9ea5-68d196c129a4","Type":"ContainerStarted","Data":"0cbffa707f9e7085f682e855db97d40ea04b839237be14772ce0809c7d4b6897"} Oct 01 09:17:21 crc kubenswrapper[4837]: I1001 09:17:21.424567 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" event={"ID":"3e10153b-fcd7-432f-95c2-c240f6b53944","Type":"ContainerStarted","Data":"e9b6edc3ea05eedaac841f8b9eaa19fed43decded9c20c18b4dec33e712955af"} Oct 01 09:17:21 crc kubenswrapper[4837]: I1001 09:17:21.461046 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" podStartSLOduration=1.666316827 podStartE2EDuration="2.460980766s" podCreationTimestamp="2025-10-01 09:17:19 +0000 UTC" firstStartedPulling="2025-10-01 09:17:20.283795279 +0000 UTC m=+7897.125402744" lastFinishedPulling="2025-10-01 09:17:21.078459218 +0000 UTC m=+7897.920066683" observedRunningTime="2025-10-01 09:17:21.453816138 +0000 UTC m=+7898.295423663" watchObservedRunningTime="2025-10-01 09:17:21.460980766 +0000 UTC m=+7898.302588241" Oct 01 09:17:21 crc kubenswrapper[4837]: I1001 09:17:21.481130 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" podStartSLOduration=2.01597031 podStartE2EDuration="2.481096194s" podCreationTimestamp="2025-10-01 09:17:19 +0000 UTC" firstStartedPulling="2025-10-01 09:17:20.193906902 +0000 UTC m=+7897.035514397" lastFinishedPulling="2025-10-01 09:17:20.659032786 +0000 UTC m=+7897.500640281" observedRunningTime="2025-10-01 09:17:21.47648859 +0000 UTC m=+7898.318096065" watchObservedRunningTime="2025-10-01 09:17:21.481096194 +0000 UTC m=+7898.322703689" Oct 01 09:17:26 crc kubenswrapper[4837]: I1001 09:17:26.817456 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:17:26 crc kubenswrapper[4837]: E1001 09:17:26.818236 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.632895 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.636088 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.659337 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.789573 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.789773 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.790024 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltx7b\" (UniqueName: \"kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.893311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.893500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltx7b\" (UniqueName: \"kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.893544 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.893972 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.894233 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.914009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltx7b\" (UniqueName: \"kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b\") pod \"redhat-operators-qkhnz\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:38 crc kubenswrapper[4837]: I1001 09:17:38.985426 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:39 crc kubenswrapper[4837]: I1001 09:17:39.508156 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:17:39 crc kubenswrapper[4837]: I1001 09:17:39.651310 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerStarted","Data":"f67fc094a5601d27171e597d01adab8c8e1182de17c4038edf8a723fa878842d"} Oct 01 09:17:40 crc kubenswrapper[4837]: I1001 09:17:40.669170 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerID="e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124" exitCode=0 Oct 01 09:17:40 crc kubenswrapper[4837]: I1001 09:17:40.669279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerDied","Data":"e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124"} Oct 01 09:17:40 crc kubenswrapper[4837]: I1001 09:17:40.672934 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:17:41 crc kubenswrapper[4837]: I1001 09:17:41.816746 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:17:41 crc kubenswrapper[4837]: E1001 09:17:41.817198 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:17:42 crc kubenswrapper[4837]: I1001 09:17:42.707945 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerStarted","Data":"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c"} Oct 01 09:17:45 crc kubenswrapper[4837]: I1001 09:17:45.751064 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerID="e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c" exitCode=0 Oct 01 09:17:45 crc kubenswrapper[4837]: I1001 09:17:45.751162 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerDied","Data":"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c"} Oct 01 09:17:46 crc kubenswrapper[4837]: I1001 09:17:46.767043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerStarted","Data":"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f"} Oct 01 09:17:46 crc kubenswrapper[4837]: I1001 09:17:46.792741 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qkhnz" podStartSLOduration=3.322742287 podStartE2EDuration="8.792674933s" podCreationTimestamp="2025-10-01 09:17:38 +0000 UTC" firstStartedPulling="2025-10-01 09:17:40.672715712 +0000 UTC m=+7917.514323167" lastFinishedPulling="2025-10-01 09:17:46.142648348 +0000 UTC m=+7922.984255813" observedRunningTime="2025-10-01 09:17:46.791686748 +0000 UTC m=+7923.633294273" watchObservedRunningTime="2025-10-01 09:17:46.792674933 +0000 UTC m=+7923.634282428" Oct 01 09:17:48 crc kubenswrapper[4837]: I1001 09:17:48.986153 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:48 crc kubenswrapper[4837]: I1001 09:17:48.986446 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:50 crc kubenswrapper[4837]: I1001 09:17:50.060744 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qkhnz" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="registry-server" probeResult="failure" output=< Oct 01 09:17:50 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:17:50 crc kubenswrapper[4837]: > Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.346357 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.349034 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.382738 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.415581 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.415667 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq9sn\" (UniqueName: \"kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.415726 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.517941 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq9sn\" (UniqueName: \"kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.518008 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.518191 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.518826 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.518857 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.541551 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq9sn\" (UniqueName: \"kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn\") pod \"redhat-marketplace-6khsj\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:53 crc kubenswrapper[4837]: I1001 09:17:53.698222 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:17:54 crc kubenswrapper[4837]: I1001 09:17:54.228978 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:17:54 crc kubenswrapper[4837]: I1001 09:17:54.865543 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerID="1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34" exitCode=0 Oct 01 09:17:54 crc kubenswrapper[4837]: I1001 09:17:54.865674 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerDied","Data":"1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34"} Oct 01 09:17:54 crc kubenswrapper[4837]: I1001 09:17:54.866061 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerStarted","Data":"46d70d4da3fbb90cfc79a1cae5bf46cffb4dc01541b098caf9808bcf1281a494"} Oct 01 09:17:56 crc kubenswrapper[4837]: I1001 09:17:56.816789 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:17:56 crc kubenswrapper[4837]: E1001 09:17:56.818162 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:17:56 crc kubenswrapper[4837]: I1001 09:17:56.899786 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerID="65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0" exitCode=0 Oct 01 09:17:56 crc kubenswrapper[4837]: I1001 09:17:56.899851 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerDied","Data":"65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0"} Oct 01 09:17:57 crc kubenswrapper[4837]: I1001 09:17:57.937142 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerStarted","Data":"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d"} Oct 01 09:17:57 crc kubenswrapper[4837]: I1001 09:17:57.967388 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6khsj" podStartSLOduration=2.347155283 podStartE2EDuration="4.967365642s" podCreationTimestamp="2025-10-01 09:17:53 +0000 UTC" firstStartedPulling="2025-10-01 09:17:54.868476563 +0000 UTC m=+7931.710084018" lastFinishedPulling="2025-10-01 09:17:57.488686892 +0000 UTC m=+7934.330294377" observedRunningTime="2025-10-01 09:17:57.961747183 +0000 UTC m=+7934.803354658" watchObservedRunningTime="2025-10-01 09:17:57.967365642 +0000 UTC m=+7934.808973097" Oct 01 09:17:59 crc kubenswrapper[4837]: I1001 09:17:59.041826 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:59 crc kubenswrapper[4837]: I1001 09:17:59.113763 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:17:59 crc kubenswrapper[4837]: I1001 09:17:59.716079 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:18:00 crc kubenswrapper[4837]: I1001 09:18:00.975086 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qkhnz" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="registry-server" containerID="cri-o://b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f" gracePeriod=2 Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.576678 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.623907 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content\") pod \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.624138 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltx7b\" (UniqueName: \"kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b\") pod \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.624480 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities\") pod \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\" (UID: \"5fe2dc8a-66f8-475c-888e-cfcf1e342140\") " Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.625776 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities" (OuterVolumeSpecName: "utilities") pod "5fe2dc8a-66f8-475c-888e-cfcf1e342140" (UID: "5fe2dc8a-66f8-475c-888e-cfcf1e342140"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.640591 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b" (OuterVolumeSpecName: "kube-api-access-ltx7b") pod "5fe2dc8a-66f8-475c-888e-cfcf1e342140" (UID: "5fe2dc8a-66f8-475c-888e-cfcf1e342140"). InnerVolumeSpecName "kube-api-access-ltx7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.722149 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe2dc8a-66f8-475c-888e-cfcf1e342140" (UID: "5fe2dc8a-66f8-475c-888e-cfcf1e342140"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.728197 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.728235 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltx7b\" (UniqueName: \"kubernetes.io/projected/5fe2dc8a-66f8-475c-888e-cfcf1e342140-kube-api-access-ltx7b\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.728249 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2dc8a-66f8-475c-888e-cfcf1e342140-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.998393 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerID="b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f" exitCode=0 Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.998529 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qkhnz" Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.999800 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerDied","Data":"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f"} Oct 01 09:18:01 crc kubenswrapper[4837]: I1001 09:18:01.999976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qkhnz" event={"ID":"5fe2dc8a-66f8-475c-888e-cfcf1e342140","Type":"ContainerDied","Data":"f67fc094a5601d27171e597d01adab8c8e1182de17c4038edf8a723fa878842d"} Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.000096 4837 scope.go:117] "RemoveContainer" containerID="b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.042267 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.051807 4837 scope.go:117] "RemoveContainer" containerID="e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.055762 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qkhnz"] Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.093325 4837 scope.go:117] "RemoveContainer" containerID="e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.164976 4837 scope.go:117] "RemoveContainer" containerID="b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f" Oct 01 09:18:02 crc kubenswrapper[4837]: E1001 09:18:02.165580 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f\": container with ID starting with b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f not found: ID does not exist" containerID="b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.165628 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f"} err="failed to get container status \"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f\": rpc error: code = NotFound desc = could not find container \"b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f\": container with ID starting with b55e47862e2ef526789c2feac677e3edee9f50a341668f7eb83f3db7afbe656f not found: ID does not exist" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.165656 4837 scope.go:117] "RemoveContainer" containerID="e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c" Oct 01 09:18:02 crc kubenswrapper[4837]: E1001 09:18:02.166403 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c\": container with ID starting with e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c not found: ID does not exist" containerID="e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.166469 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c"} err="failed to get container status \"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c\": rpc error: code = NotFound desc = could not find container \"e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c\": container with ID starting with e8225a7c2ec552eb4026fc4d9a9e1406731cc79cb8bd57f42650ec4e5da1b27c not found: ID does not exist" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.166513 4837 scope.go:117] "RemoveContainer" containerID="e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124" Oct 01 09:18:02 crc kubenswrapper[4837]: E1001 09:18:02.167289 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124\": container with ID starting with e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124 not found: ID does not exist" containerID="e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124" Oct 01 09:18:02 crc kubenswrapper[4837]: I1001 09:18:02.167353 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124"} err="failed to get container status \"e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124\": rpc error: code = NotFound desc = could not find container \"e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124\": container with ID starting with e808cf16056b032223793d97689266ccbfa9655d3bf493448bee5b8793a9d124 not found: ID does not exist" Oct 01 09:18:03 crc kubenswrapper[4837]: I1001 09:18:03.698560 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:03 crc kubenswrapper[4837]: I1001 09:18:03.698918 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:03 crc kubenswrapper[4837]: I1001 09:18:03.748758 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:03 crc kubenswrapper[4837]: I1001 09:18:03.831767 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" path="/var/lib/kubelet/pods/5fe2dc8a-66f8-475c-888e-cfcf1e342140/volumes" Oct 01 09:18:04 crc kubenswrapper[4837]: I1001 09:18:04.096475 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:05 crc kubenswrapper[4837]: I1001 09:18:05.117169 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.065507 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6khsj" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="registry-server" containerID="cri-o://5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d" gracePeriod=2 Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.071934 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-qjgjc"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.084363 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2wmlh"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.100745 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2wmlh"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.110923 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-qjgjc"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.119196 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8n975"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.129340 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8n975"] Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.652395 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.752014 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq9sn\" (UniqueName: \"kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn\") pod \"6b76420f-03df-49e4-8579-e62bf16b82a4\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.752640 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities\") pod \"6b76420f-03df-49e4-8579-e62bf16b82a4\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.753112 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content\") pod \"6b76420f-03df-49e4-8579-e62bf16b82a4\" (UID: \"6b76420f-03df-49e4-8579-e62bf16b82a4\") " Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.753397 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities" (OuterVolumeSpecName: "utilities") pod "6b76420f-03df-49e4-8579-e62bf16b82a4" (UID: "6b76420f-03df-49e4-8579-e62bf16b82a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.758368 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn" (OuterVolumeSpecName: "kube-api-access-zq9sn") pod "6b76420f-03df-49e4-8579-e62bf16b82a4" (UID: "6b76420f-03df-49e4-8579-e62bf16b82a4"). InnerVolumeSpecName "kube-api-access-zq9sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.764362 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq9sn\" (UniqueName: \"kubernetes.io/projected/6b76420f-03df-49e4-8579-e62bf16b82a4-kube-api-access-zq9sn\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.764398 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.771284 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b76420f-03df-49e4-8579-e62bf16b82a4" (UID: "6b76420f-03df-49e4-8579-e62bf16b82a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:18:06 crc kubenswrapper[4837]: I1001 09:18:06.866092 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b76420f-03df-49e4-8579-e62bf16b82a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.087378 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerID="5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d" exitCode=0 Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.087465 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerDied","Data":"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d"} Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.087514 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6khsj" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.087538 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6khsj" event={"ID":"6b76420f-03df-49e4-8579-e62bf16b82a4","Type":"ContainerDied","Data":"46d70d4da3fbb90cfc79a1cae5bf46cffb4dc01541b098caf9808bcf1281a494"} Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.087569 4837 scope.go:117] "RemoveContainer" containerID="5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.132032 4837 scope.go:117] "RemoveContainer" containerID="65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.151538 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.158653 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6khsj"] Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.178101 4837 scope.go:117] "RemoveContainer" containerID="1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.215301 4837 scope.go:117] "RemoveContainer" containerID="5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d" Oct 01 09:18:07 crc kubenswrapper[4837]: E1001 09:18:07.216081 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d\": container with ID starting with 5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d not found: ID does not exist" containerID="5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.216149 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d"} err="failed to get container status \"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d\": rpc error: code = NotFound desc = could not find container \"5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d\": container with ID starting with 5ec4b925ce9835c0e42877812f452e912d160d4364e660f680efee3552346a0d not found: ID does not exist" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.216191 4837 scope.go:117] "RemoveContainer" containerID="65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0" Oct 01 09:18:07 crc kubenswrapper[4837]: E1001 09:18:07.216711 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0\": container with ID starting with 65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0 not found: ID does not exist" containerID="65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.216747 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0"} err="failed to get container status \"65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0\": rpc error: code = NotFound desc = could not find container \"65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0\": container with ID starting with 65f3789ebd94d3653ab9d750aca7d19552d92a65136be80c0ff81bc5614c25e0 not found: ID does not exist" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.216777 4837 scope.go:117] "RemoveContainer" containerID="1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34" Oct 01 09:18:07 crc kubenswrapper[4837]: E1001 09:18:07.217200 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34\": container with ID starting with 1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34 not found: ID does not exist" containerID="1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.217234 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34"} err="failed to get container status \"1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34\": rpc error: code = NotFound desc = could not find container \"1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34\": container with ID starting with 1108ee796c067c3f99300f40d31e0a015786193a964c3133e8e92898981ceb34 not found: ID does not exist" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.826736 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" path="/var/lib/kubelet/pods/6b76420f-03df-49e4-8579-e62bf16b82a4/volumes" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.827579 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1095f7c-8690-4423-a608-0aeca0c1871c" path="/var/lib/kubelet/pods/c1095f7c-8690-4423-a608-0aeca0c1871c/volumes" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.828078 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e6452d-fee6-43ca-9845-43f6f30c2a29" path="/var/lib/kubelet/pods/e0e6452d-fee6-43ca-9845-43f6f30c2a29/volumes" Oct 01 09:18:07 crc kubenswrapper[4837]: I1001 09:18:07.828989 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7403bed-3c82-49c3-8d57-75470652e3db" path="/var/lib/kubelet/pods/e7403bed-3c82-49c3-8d57-75470652e3db/volumes" Oct 01 09:18:08 crc kubenswrapper[4837]: I1001 09:18:08.816025 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:18:08 crc kubenswrapper[4837]: E1001 09:18:08.816490 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:18:13 crc kubenswrapper[4837]: I1001 09:18:13.666235 4837 scope.go:117] "RemoveContainer" containerID="ae0488ee5ab012d506b74c065262a89f049462c1051efc738e936b5f1ba33a79" Oct 01 09:18:13 crc kubenswrapper[4837]: I1001 09:18:13.711606 4837 scope.go:117] "RemoveContainer" containerID="c7c70daef7196a6c174366bf118012c84eb95952bb75c36be847e684a72cae9c" Oct 01 09:18:13 crc kubenswrapper[4837]: I1001 09:18:13.758995 4837 scope.go:117] "RemoveContainer" containerID="cd96e17d423b561dc071a5ced945d6261327d8ca80de9dde1e9d81b38add842e" Oct 01 09:18:16 crc kubenswrapper[4837]: I1001 09:18:16.027735 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2ea8-account-create-p2hkr"] Oct 01 09:18:16 crc kubenswrapper[4837]: I1001 09:18:16.036041 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2ea8-account-create-p2hkr"] Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.043773 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3a18-account-create-qfkdm"] Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.055267 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cefb-account-create-6mgpb"] Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.064486 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3a18-account-create-qfkdm"] Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.074172 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cefb-account-create-6mgpb"] Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.835096 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036173d9-f63f-4124-9278-594a6f412eb6" path="/var/lib/kubelet/pods/036173d9-f63f-4124-9278-594a6f412eb6/volumes" Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.835970 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd8a385-1c4d-4878-853d-3140eaf1fbd8" path="/var/lib/kubelet/pods/9dd8a385-1c4d-4878-853d-3140eaf1fbd8/volumes" Oct 01 09:18:17 crc kubenswrapper[4837]: I1001 09:18:17.836560 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2fad0f2-8515-4a39-96b4-adf4faf14545" path="/var/lib/kubelet/pods/c2fad0f2-8515-4a39-96b4-adf4faf14545/volumes" Oct 01 09:18:20 crc kubenswrapper[4837]: I1001 09:18:20.816716 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:18:20 crc kubenswrapper[4837]: E1001 09:18:20.817465 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:18:31 crc kubenswrapper[4837]: I1001 09:18:31.487906 4837 patch_prober.go:28] interesting pod/oauth-openshift-6cc7c68bbf-pwxfl container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.54:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 09:18:31 crc kubenswrapper[4837]: I1001 09:18:31.488730 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-6cc7c68bbf-pwxfl" podUID="7c9503d8-6366-4d81-afb2-029f554bebe4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.54:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:18:33 crc kubenswrapper[4837]: I1001 09:18:33.823072 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:18:33 crc kubenswrapper[4837]: E1001 09:18:33.823767 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:18:34 crc kubenswrapper[4837]: I1001 09:18:34.049539 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qlgcs"] Oct 01 09:18:34 crc kubenswrapper[4837]: I1001 09:18:34.057304 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-qlgcs"] Oct 01 09:18:35 crc kubenswrapper[4837]: I1001 09:18:35.831162 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a51e260-c980-4c8d-94d3-cd1f8e9c6f66" path="/var/lib/kubelet/pods/1a51e260-c980-4c8d-94d3-cd1f8e9c6f66/volumes" Oct 01 09:18:45 crc kubenswrapper[4837]: I1001 09:18:45.816443 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:18:45 crc kubenswrapper[4837]: E1001 09:18:45.817428 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:18:48 crc kubenswrapper[4837]: I1001 09:18:48.059867 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ktt7n"] Oct 01 09:18:48 crc kubenswrapper[4837]: I1001 09:18:48.073936 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ktt7n"] Oct 01 09:18:49 crc kubenswrapper[4837]: I1001 09:18:49.832960 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33918084-d6ec-433c-9fbe-8f1386fb347e" path="/var/lib/kubelet/pods/33918084-d6ec-433c-9fbe-8f1386fb347e/volumes" Oct 01 09:18:50 crc kubenswrapper[4837]: I1001 09:18:50.032058 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pj4qg"] Oct 01 09:18:50 crc kubenswrapper[4837]: I1001 09:18:50.042723 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pj4qg"] Oct 01 09:18:51 crc kubenswrapper[4837]: I1001 09:18:51.841392 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cafbdbfe-2005-402e-a731-20f4afb49879" path="/var/lib/kubelet/pods/cafbdbfe-2005-402e-a731-20f4afb49879/volumes" Oct 01 09:18:56 crc kubenswrapper[4837]: I1001 09:18:56.816102 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:18:56 crc kubenswrapper[4837]: E1001 09:18:56.816859 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:19:08 crc kubenswrapper[4837]: I1001 09:19:08.816063 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:19:08 crc kubenswrapper[4837]: E1001 09:19:08.817115 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:19:09 crc kubenswrapper[4837]: I1001 09:19:09.046678 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9n5g"] Oct 01 09:19:09 crc kubenswrapper[4837]: I1001 09:19:09.063763 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-t9n5g"] Oct 01 09:19:09 crc kubenswrapper[4837]: I1001 09:19:09.831414 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b63f46a7-38b3-4529-bbc2-ea5665df205f" path="/var/lib/kubelet/pods/b63f46a7-38b3-4529-bbc2-ea5665df205f/volumes" Oct 01 09:19:13 crc kubenswrapper[4837]: I1001 09:19:13.948566 4837 scope.go:117] "RemoveContainer" containerID="575d3a2180b5688693aec0ce3cc809d3baa4495b5243b96de56e531277ef64b1" Oct 01 09:19:13 crc kubenswrapper[4837]: I1001 09:19:13.991330 4837 scope.go:117] "RemoveContainer" containerID="62791ee7dccb440098f077245e147c6f69d5319ed6da30e8620e5bb40529d030" Oct 01 09:19:14 crc kubenswrapper[4837]: I1001 09:19:14.090577 4837 scope.go:117] "RemoveContainer" containerID="cf80aa3de92278f5af049dcc2a1d5b6ff02a73f47367c24496960c65cf828ee7" Oct 01 09:19:14 crc kubenswrapper[4837]: I1001 09:19:14.139163 4837 scope.go:117] "RemoveContainer" containerID="b2a2a1c1bcc52ffef6d39b999104801f7a2c7edc1123eb6d4388d0ee634f4564" Oct 01 09:19:14 crc kubenswrapper[4837]: I1001 09:19:14.159113 4837 scope.go:117] "RemoveContainer" containerID="9e2d383f5619fa58b7348c2c88ef6434a049bfb1e28579cb4884eb0279ddae32" Oct 01 09:19:14 crc kubenswrapper[4837]: I1001 09:19:14.239370 4837 scope.go:117] "RemoveContainer" containerID="1d2102928b6c3afb9e9f1d360c6b17f3730feabe302ad315b3cf53004b32ba8f" Oct 01 09:19:14 crc kubenswrapper[4837]: I1001 09:19:14.270636 4837 scope.go:117] "RemoveContainer" containerID="1345222af4a5403f43eb2a1848d36f26b54d28ca266554c5443a182034f59ec2" Oct 01 09:19:19 crc kubenswrapper[4837]: I1001 09:19:19.817068 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:19:19 crc kubenswrapper[4837]: E1001 09:19:19.817665 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:19:31 crc kubenswrapper[4837]: I1001 09:19:31.816506 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:19:33 crc kubenswrapper[4837]: I1001 09:19:33.068052 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698"} Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.363946 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365107 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="extract-utilities" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365125 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="extract-utilities" Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365154 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365163 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365190 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="extract-utilities" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365199 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="extract-utilities" Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365214 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="extract-content" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365221 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="extract-content" Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365239 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="extract-content" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365247 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="extract-content" Oct 01 09:21:33 crc kubenswrapper[4837]: E1001 09:21:33.365264 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365271 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365505 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b76420f-03df-49e4-8579-e62bf16b82a4" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.365539 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe2dc8a-66f8-475c-888e-cfcf1e342140" containerName="registry-server" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.367584 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.391427 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.484874 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.485009 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqc74\" (UniqueName: \"kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.485166 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.587216 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.587274 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqc74\" (UniqueName: \"kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.587346 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.588164 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.588577 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.606372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqc74\" (UniqueName: \"kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74\") pod \"community-operators-56zph\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:33 crc kubenswrapper[4837]: I1001 09:21:33.703195 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:34 crc kubenswrapper[4837]: I1001 09:21:34.250331 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:34 crc kubenswrapper[4837]: I1001 09:21:34.543566 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerID="e06df847d18f96f6dfa79def67a92183be9c94d8ef96aeaa6145792e50041898" exitCode=0 Oct 01 09:21:34 crc kubenswrapper[4837]: I1001 09:21:34.543643 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerDied","Data":"e06df847d18f96f6dfa79def67a92183be9c94d8ef96aeaa6145792e50041898"} Oct 01 09:21:34 crc kubenswrapper[4837]: I1001 09:21:34.543897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerStarted","Data":"addcc71aa04ecc02f0c9b3f08a366c151d04cc8cb91e0b33b3e4e58c75339399"} Oct 01 09:21:35 crc kubenswrapper[4837]: I1001 09:21:35.557850 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerStarted","Data":"56e65d1eff4a3636036e5920f7edfe7e6121f27821ccb5c9b828b491d74ff0b6"} Oct 01 09:21:37 crc kubenswrapper[4837]: I1001 09:21:37.582409 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerID="56e65d1eff4a3636036e5920f7edfe7e6121f27821ccb5c9b828b491d74ff0b6" exitCode=0 Oct 01 09:21:37 crc kubenswrapper[4837]: I1001 09:21:37.582508 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerDied","Data":"56e65d1eff4a3636036e5920f7edfe7e6121f27821ccb5c9b828b491d74ff0b6"} Oct 01 09:21:37 crc kubenswrapper[4837]: I1001 09:21:37.591837 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fc52d12-9bf2-4649-9ea5-68d196c129a4" containerID="0cbffa707f9e7085f682e855db97d40ea04b839237be14772ce0809c7d4b6897" exitCode=0 Oct 01 09:21:37 crc kubenswrapper[4837]: I1001 09:21:37.591907 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" event={"ID":"5fc52d12-9bf2-4649-9ea5-68d196c129a4","Type":"ContainerDied","Data":"0cbffa707f9e7085f682e855db97d40ea04b839237be14772ce0809c7d4b6897"} Oct 01 09:21:38 crc kubenswrapper[4837]: I1001 09:21:38.611783 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerStarted","Data":"2f5dce9929f2b7c5ad55f0736e523c0a87988b7bef01ec619f5fab19f74fe05d"} Oct 01 09:21:38 crc kubenswrapper[4837]: I1001 09:21:38.641286 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-56zph" podStartSLOduration=2.196830054 podStartE2EDuration="5.641260385s" podCreationTimestamp="2025-10-01 09:21:33 +0000 UTC" firstStartedPulling="2025-10-01 09:21:34.546083761 +0000 UTC m=+8151.387691226" lastFinishedPulling="2025-10-01 09:21:37.990514102 +0000 UTC m=+8154.832121557" observedRunningTime="2025-10-01 09:21:38.637857001 +0000 UTC m=+8155.479464476" watchObservedRunningTime="2025-10-01 09:21:38.641260385 +0000 UTC m=+8155.482867850" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.087818 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.210203 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key\") pod \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.210313 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory\") pod \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.210525 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pwn8\" (UniqueName: \"kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8\") pod \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.210625 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle\") pod \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\" (UID: \"5fc52d12-9bf2-4649-9ea5-68d196c129a4\") " Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.218342 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8" (OuterVolumeSpecName: "kube-api-access-6pwn8") pod "5fc52d12-9bf2-4649-9ea5-68d196c129a4" (UID: "5fc52d12-9bf2-4649-9ea5-68d196c129a4"). InnerVolumeSpecName "kube-api-access-6pwn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.218442 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "5fc52d12-9bf2-4649-9ea5-68d196c129a4" (UID: "5fc52d12-9bf2-4649-9ea5-68d196c129a4"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.259049 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fc52d12-9bf2-4649-9ea5-68d196c129a4" (UID: "5fc52d12-9bf2-4649-9ea5-68d196c129a4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.261011 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory" (OuterVolumeSpecName: "inventory") pod "5fc52d12-9bf2-4649-9ea5-68d196c129a4" (UID: "5fc52d12-9bf2-4649-9ea5-68d196c129a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.312616 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pwn8\" (UniqueName: \"kubernetes.io/projected/5fc52d12-9bf2-4649-9ea5-68d196c129a4-kube-api-access-6pwn8\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.312656 4837 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.312666 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.312675 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc52d12-9bf2-4649-9ea5-68d196c129a4-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.635193 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" event={"ID":"5fc52d12-9bf2-4649-9ea5-68d196c129a4","Type":"ContainerDied","Data":"de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508"} Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.635263 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de5965530a600e390b18fb4666bdc5d4f7a371373245c1bb66574e273083a508" Oct 01 09:21:39 crc kubenswrapper[4837]: I1001 09:21:39.635260 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm" Oct 01 09:21:43 crc kubenswrapper[4837]: I1001 09:21:43.703865 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:43 crc kubenswrapper[4837]: I1001 09:21:43.704330 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:43 crc kubenswrapper[4837]: I1001 09:21:43.756233 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:44 crc kubenswrapper[4837]: I1001 09:21:44.753264 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:44 crc kubenswrapper[4837]: I1001 09:21:44.811741 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:46 crc kubenswrapper[4837]: I1001 09:21:46.714951 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-56zph" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="registry-server" containerID="cri-o://2f5dce9929f2b7c5ad55f0736e523c0a87988b7bef01ec619f5fab19f74fe05d" gracePeriod=2 Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.726653 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerID="2f5dce9929f2b7c5ad55f0736e523c0a87988b7bef01ec619f5fab19f74fe05d" exitCode=0 Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.726730 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerDied","Data":"2f5dce9929f2b7c5ad55f0736e523c0a87988b7bef01ec619f5fab19f74fe05d"} Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.727340 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56zph" event={"ID":"3e1a2576-c96c-427f-928f-650d5e7f0d9c","Type":"ContainerDied","Data":"addcc71aa04ecc02f0c9b3f08a366c151d04cc8cb91e0b33b3e4e58c75339399"} Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.727359 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="addcc71aa04ecc02f0c9b3f08a366c151d04cc8cb91e0b33b3e4e58c75339399" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.749115 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.802767 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities\") pod \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.802900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content\") pod \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.803186 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqc74\" (UniqueName: \"kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74\") pod \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\" (UID: \"3e1a2576-c96c-427f-928f-650d5e7f0d9c\") " Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.803793 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities" (OuterVolumeSpecName: "utilities") pod "3e1a2576-c96c-427f-928f-650d5e7f0d9c" (UID: "3e1a2576-c96c-427f-928f-650d5e7f0d9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.808062 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74" (OuterVolumeSpecName: "kube-api-access-mqc74") pod "3e1a2576-c96c-427f-928f-650d5e7f0d9c" (UID: "3e1a2576-c96c-427f-928f-650d5e7f0d9c"). InnerVolumeSpecName "kube-api-access-mqc74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.858895 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e1a2576-c96c-427f-928f-650d5e7f0d9c" (UID: "3e1a2576-c96c-427f-928f-650d5e7f0d9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.906305 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqc74\" (UniqueName: \"kubernetes.io/projected/3e1a2576-c96c-427f-928f-650d5e7f0d9c-kube-api-access-mqc74\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.906344 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:47 crc kubenswrapper[4837]: I1001 09:21:47.906357 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e1a2576-c96c-427f-928f-650d5e7f0d9c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:21:48 crc kubenswrapper[4837]: I1001 09:21:48.738503 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56zph" Oct 01 09:21:48 crc kubenswrapper[4837]: I1001 09:21:48.791749 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:48 crc kubenswrapper[4837]: I1001 09:21:48.806884 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-56zph"] Oct 01 09:21:49 crc kubenswrapper[4837]: I1001 09:21:49.832495 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" path="/var/lib/kubelet/pods/3e1a2576-c96c-427f-928f-650d5e7f0d9c/volumes" Oct 01 09:21:53 crc kubenswrapper[4837]: I1001 09:21:53.079154 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:21:53 crc kubenswrapper[4837]: I1001 09:21:53.079539 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:22:11 crc kubenswrapper[4837]: I1001 09:22:11.057055 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-fgb2r"] Oct 01 09:22:11 crc kubenswrapper[4837]: I1001 09:22:11.067261 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-fgb2r"] Oct 01 09:22:11 crc kubenswrapper[4837]: I1001 09:22:11.835010 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33c79e5a-fb36-4447-a489-4a0493d1e56f" path="/var/lib/kubelet/pods/33c79e5a-fb36-4447-a489-4a0493d1e56f/volumes" Oct 01 09:22:14 crc kubenswrapper[4837]: I1001 09:22:14.517435 4837 scope.go:117] "RemoveContainer" containerID="43a533c97781838f30f31464ecaa2bc7ea99293b8e338de45539c8a6798ccd9b" Oct 01 09:22:14 crc kubenswrapper[4837]: I1001 09:22:14.552138 4837 scope.go:117] "RemoveContainer" containerID="a8462774cf7a801136982b3a4b5c916ba1c45069d9a6874c59d2fa526584b18c" Oct 01 09:22:14 crc kubenswrapper[4837]: I1001 09:22:14.582504 4837 scope.go:117] "RemoveContainer" containerID="584b256413d099bc484ef115065bad072f46a82b7b9633f0218f67faf4d2e53a" Oct 01 09:22:14 crc kubenswrapper[4837]: I1001 09:22:14.604375 4837 scope.go:117] "RemoveContainer" containerID="574204b2ed6f85400d2db6410c3d21c66a0f961beb34b1c9994af10b8473d4a4" Oct 01 09:22:14 crc kubenswrapper[4837]: I1001 09:22:14.630631 4837 scope.go:117] "RemoveContainer" containerID="91582e7ecb939bb8896ec2ec40a8583209bc646b3e57287e25b9c79db07672e9" Oct 01 09:22:21 crc kubenswrapper[4837]: I1001 09:22:21.035955 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-0856-account-create-5pgsg"] Oct 01 09:22:21 crc kubenswrapper[4837]: I1001 09:22:21.054630 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-0856-account-create-5pgsg"] Oct 01 09:22:21 crc kubenswrapper[4837]: I1001 09:22:21.836592 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc46596a-3455-475b-8f07-4c66ac27b9c6" path="/var/lib/kubelet/pods/fc46596a-3455-475b-8f07-4c66ac27b9c6/volumes" Oct 01 09:22:23 crc kubenswrapper[4837]: I1001 09:22:23.079014 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:22:23 crc kubenswrapper[4837]: I1001 09:22:23.079315 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:22:36 crc kubenswrapper[4837]: I1001 09:22:36.063626 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-56ckh"] Oct 01 09:22:36 crc kubenswrapper[4837]: I1001 09:22:36.074136 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-56ckh"] Oct 01 09:22:37 crc kubenswrapper[4837]: I1001 09:22:37.827064 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5561324-9c39-4292-80a9-cad1343b9d87" path="/var/lib/kubelet/pods/d5561324-9c39-4292-80a9-cad1343b9d87/volumes" Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.079307 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.080094 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.080175 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.081284 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.081428 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698" gracePeriod=600 Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.468511 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698" exitCode=0 Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.468638 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698"} Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.468999 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b"} Oct 01 09:22:53 crc kubenswrapper[4837]: I1001 09:22:53.469024 4837 scope.go:117] "RemoveContainer" containerID="6068fd05b813359d25a888ae0dcff41f76af59fb05df3fd6f5e772bfd5f648b5" Oct 01 09:23:14 crc kubenswrapper[4837]: I1001 09:23:14.722596 4837 scope.go:117] "RemoveContainer" containerID="2640c611abcb31747b8eaf066fa23193072bbccfd334f1e7b791ded21f1a5aa7" Oct 01 09:23:14 crc kubenswrapper[4837]: I1001 09:23:14.749946 4837 scope.go:117] "RemoveContainer" containerID="394f6af21f4b50354f4acbd0852186864979bd7858f7fa1f11441792548e77dd" Oct 01 09:23:14 crc kubenswrapper[4837]: I1001 09:23:14.839004 4837 scope.go:117] "RemoveContainer" containerID="a7f2291be7d22a76fba9a3d959ddc6ef518dcf253635ac4bbde334cd4db9fc6f" Oct 01 09:23:14 crc kubenswrapper[4837]: I1001 09:23:14.883572 4837 scope.go:117] "RemoveContainer" containerID="8d779aec4868120c882adc08c1c873dc2853ba3fe1f7e05630a935c1d9989a7c" Oct 01 09:23:14 crc kubenswrapper[4837]: I1001 09:23:14.929342 4837 scope.go:117] "RemoveContainer" containerID="58c6b92f330f632a64066c14383d9ce1b675787311bef343511c77e7df449a93" Oct 01 09:24:53 crc kubenswrapper[4837]: I1001 09:24:53.079903 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:24:53 crc kubenswrapper[4837]: I1001 09:24:53.080676 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:25:00 crc kubenswrapper[4837]: I1001 09:25:00.068169 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-26rgd"] Oct 01 09:25:00 crc kubenswrapper[4837]: I1001 09:25:00.085164 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-26rgd"] Oct 01 09:25:01 crc kubenswrapper[4837]: I1001 09:25:01.828907 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="890300f8-e949-4add-9103-3ddab9b0de07" path="/var/lib/kubelet/pods/890300f8-e949-4add-9103-3ddab9b0de07/volumes" Oct 01 09:25:10 crc kubenswrapper[4837]: I1001 09:25:10.063775 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-79c0-account-create-gvrk7"] Oct 01 09:25:10 crc kubenswrapper[4837]: I1001 09:25:10.075408 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-79c0-account-create-gvrk7"] Oct 01 09:25:11 crc kubenswrapper[4837]: I1001 09:25:11.835266 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b035865-59b0-46f5-ae64-877c1ce99e05" path="/var/lib/kubelet/pods/2b035865-59b0-46f5-ae64-877c1ce99e05/volumes" Oct 01 09:25:15 crc kubenswrapper[4837]: I1001 09:25:15.092027 4837 scope.go:117] "RemoveContainer" containerID="36e8a3c24a1ff9c9c4225b06a40a4cea653c78ee753ae9cded7d0886645d48b8" Oct 01 09:25:15 crc kubenswrapper[4837]: I1001 09:25:15.128367 4837 scope.go:117] "RemoveContainer" containerID="d1e73b6a3db5f5773c31234ea6dfd7fecc50495c288ddfd8f49523bb89237f14" Oct 01 09:25:20 crc kubenswrapper[4837]: I1001 09:25:20.265394 4837 generic.go:334] "Generic (PLEG): container finished" podID="3e10153b-fcd7-432f-95c2-c240f6b53944" containerID="e9b6edc3ea05eedaac841f8b9eaa19fed43decded9c20c18b4dec33e712955af" exitCode=0 Oct 01 09:25:20 crc kubenswrapper[4837]: I1001 09:25:20.265477 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" event={"ID":"3e10153b-fcd7-432f-95c2-c240f6b53944","Type":"ContainerDied","Data":"e9b6edc3ea05eedaac841f8b9eaa19fed43decded9c20c18b4dec33e712955af"} Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.796336 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.898459 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle\") pod \"3e10153b-fcd7-432f-95c2-c240f6b53944\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.898528 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key\") pod \"3e10153b-fcd7-432f-95c2-c240f6b53944\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.898833 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btmhg\" (UniqueName: \"kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg\") pod \"3e10153b-fcd7-432f-95c2-c240f6b53944\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.898866 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory\") pod \"3e10153b-fcd7-432f-95c2-c240f6b53944\" (UID: \"3e10153b-fcd7-432f-95c2-c240f6b53944\") " Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.905435 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg" (OuterVolumeSpecName: "kube-api-access-btmhg") pod "3e10153b-fcd7-432f-95c2-c240f6b53944" (UID: "3e10153b-fcd7-432f-95c2-c240f6b53944"). InnerVolumeSpecName "kube-api-access-btmhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.905480 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "3e10153b-fcd7-432f-95c2-c240f6b53944" (UID: "3e10153b-fcd7-432f-95c2-c240f6b53944"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.933254 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3e10153b-fcd7-432f-95c2-c240f6b53944" (UID: "3e10153b-fcd7-432f-95c2-c240f6b53944"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:25:21 crc kubenswrapper[4837]: I1001 09:25:21.955719 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory" (OuterVolumeSpecName: "inventory") pod "3e10153b-fcd7-432f-95c2-c240f6b53944" (UID: "3e10153b-fcd7-432f-95c2-c240f6b53944"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.002395 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btmhg\" (UniqueName: \"kubernetes.io/projected/3e10153b-fcd7-432f-95c2-c240f6b53944-kube-api-access-btmhg\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.002465 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.002495 4837 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.002523 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e10153b-fcd7-432f-95c2-c240f6b53944-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.292138 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" event={"ID":"3e10153b-fcd7-432f-95c2-c240f6b53944","Type":"ContainerDied","Data":"a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa"} Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.292424 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8516cdf2060fec5f1566d516d9d91e5d87c545b756012e12d74e8bdca33abfa" Oct 01 09:25:22 crc kubenswrapper[4837]: I1001 09:25:22.292261 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t" Oct 01 09:25:23 crc kubenswrapper[4837]: I1001 09:25:23.046657 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-6fg5l"] Oct 01 09:25:23 crc kubenswrapper[4837]: I1001 09:25:23.064676 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-6fg5l"] Oct 01 09:25:23 crc kubenswrapper[4837]: I1001 09:25:23.080631 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:25:23 crc kubenswrapper[4837]: I1001 09:25:23.080791 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:25:23 crc kubenswrapper[4837]: I1001 09:25:23.832067 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6fa401e-5355-41db-8257-e0f97eab28de" path="/var/lib/kubelet/pods/c6fa401e-5355-41db-8257-e0f97eab28de/volumes" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.505558 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lhpww"] Oct 01 09:25:47 crc kubenswrapper[4837]: E1001 09:25:47.507426 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e10153b-fcd7-432f-95c2-c240f6b53944" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.507468 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e10153b-fcd7-432f-95c2-c240f6b53944" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 01 09:25:47 crc kubenswrapper[4837]: E1001 09:25:47.507516 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="extract-utilities" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.507536 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="extract-utilities" Oct 01 09:25:47 crc kubenswrapper[4837]: E1001 09:25:47.507569 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="registry-server" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.507589 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="registry-server" Oct 01 09:25:47 crc kubenswrapper[4837]: E1001 09:25:47.507633 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="extract-content" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.507652 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="extract-content" Oct 01 09:25:47 crc kubenswrapper[4837]: E1001 09:25:47.507738 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc52d12-9bf2-4649-9ea5-68d196c129a4" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.507763 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc52d12-9bf2-4649-9ea5-68d196c129a4" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.508308 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc52d12-9bf2-4649-9ea5-68d196c129a4" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.508364 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e10153b-fcd7-432f-95c2-c240f6b53944" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.508414 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1a2576-c96c-427f-928f-650d5e7f0d9c" containerName="registry-server" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.510228 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.518072 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.518470 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.518577 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.519156 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.525340 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-trbtg"] Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.527145 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.529448 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.531535 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.549766 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lhpww"] Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.566945 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-trbtg"] Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.569319 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.569390 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.569475 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.570589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7phr7\" (UniqueName: \"kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672006 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672084 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672118 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672142 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672202 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7phr7\" (UniqueName: \"kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672222 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lgnv\" (UniqueName: \"kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.672390 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.677393 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.683169 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.684072 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.693136 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7phr7\" (UniqueName: \"kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7\") pod \"bootstrap-openstack-openstack-cell1-lhpww\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.775951 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.776158 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.776287 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lgnv\" (UniqueName: \"kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.776343 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.780622 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.781137 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.783484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.799927 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lgnv\" (UniqueName: \"kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv\") pod \"bootstrap-openstack-openstack-networker-trbtg\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.830369 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:25:47 crc kubenswrapper[4837]: I1001 09:25:47.847868 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:25:48 crc kubenswrapper[4837]: I1001 09:25:48.457422 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:25:48 crc kubenswrapper[4837]: I1001 09:25:48.488769 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-lhpww"] Oct 01 09:25:48 crc kubenswrapper[4837]: I1001 09:25:48.530253 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-trbtg"] Oct 01 09:25:48 crc kubenswrapper[4837]: W1001 09:25:48.535141 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b2e57be_af14_4e5b_8751_dc110da24b25.slice/crio-f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0 WatchSource:0}: Error finding container f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0: Status 404 returned error can't find the container with id f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0 Oct 01 09:25:48 crc kubenswrapper[4837]: I1001 09:25:48.566154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" event={"ID":"7b2e57be-af14-4e5b-8751-dc110da24b25","Type":"ContainerStarted","Data":"f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0"} Oct 01 09:25:48 crc kubenswrapper[4837]: I1001 09:25:48.567430 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" event={"ID":"18443fd2-82a7-4be3-9ad7-8be4777ec6e5","Type":"ContainerStarted","Data":"6b8568d8c3dcf3c2b3d2ebfb2857851b478066c2371af12bec6591cc2f58df84"} Oct 01 09:25:49 crc kubenswrapper[4837]: I1001 09:25:49.586920 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" event={"ID":"7b2e57be-af14-4e5b-8751-dc110da24b25","Type":"ContainerStarted","Data":"e31bc9d0a521d4afe5881cfa01f9506b8158365e64a2751af1af9330c97b3a76"} Oct 01 09:25:49 crc kubenswrapper[4837]: I1001 09:25:49.588684 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" event={"ID":"18443fd2-82a7-4be3-9ad7-8be4777ec6e5","Type":"ContainerStarted","Data":"fd291b579eb3ff0e0ac865fb871de393dfdf22480ec9a5667ea4b7ad9659d9b4"} Oct 01 09:25:49 crc kubenswrapper[4837]: I1001 09:25:49.616829 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" podStartSLOduration=2.070264166 podStartE2EDuration="2.616807868s" podCreationTimestamp="2025-10-01 09:25:47 +0000 UTC" firstStartedPulling="2025-10-01 09:25:48.537495242 +0000 UTC m=+8405.379102697" lastFinishedPulling="2025-10-01 09:25:49.084038914 +0000 UTC m=+8405.925646399" observedRunningTime="2025-10-01 09:25:49.601412759 +0000 UTC m=+8406.443020224" watchObservedRunningTime="2025-10-01 09:25:49.616807868 +0000 UTC m=+8406.458415343" Oct 01 09:25:49 crc kubenswrapper[4837]: I1001 09:25:49.638333 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" podStartSLOduration=2.161900754 podStartE2EDuration="2.638314908s" podCreationTimestamp="2025-10-01 09:25:47 +0000 UTC" firstStartedPulling="2025-10-01 09:25:48.4570929 +0000 UTC m=+8405.298700355" lastFinishedPulling="2025-10-01 09:25:48.933507054 +0000 UTC m=+8405.775114509" observedRunningTime="2025-10-01 09:25:49.630139957 +0000 UTC m=+8406.471747452" watchObservedRunningTime="2025-10-01 09:25:49.638314908 +0000 UTC m=+8406.479922363" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.079264 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.079939 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.080026 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.081222 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.081329 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" gracePeriod=600 Oct 01 09:25:53 crc kubenswrapper[4837]: E1001 09:25:53.207892 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.637869 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" exitCode=0 Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.637960 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b"} Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.638171 4837 scope.go:117] "RemoveContainer" containerID="465c400745692616b1104fd3bab7b7349a89bd6194c3e9b3c84f0cb6c4e75698" Oct 01 09:25:53 crc kubenswrapper[4837]: I1001 09:25:53.638884 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:25:53 crc kubenswrapper[4837]: E1001 09:25:53.639210 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:26:04 crc kubenswrapper[4837]: I1001 09:26:04.818096 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:26:04 crc kubenswrapper[4837]: E1001 09:26:04.819204 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:26:15 crc kubenswrapper[4837]: I1001 09:26:15.223813 4837 scope.go:117] "RemoveContainer" containerID="c213a609b3f454910eaa1fb8ca621dfbec8fd94396bb083b312a832241c19640" Oct 01 09:26:19 crc kubenswrapper[4837]: I1001 09:26:19.816166 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:26:19 crc kubenswrapper[4837]: E1001 09:26:19.817102 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:26:33 crc kubenswrapper[4837]: I1001 09:26:33.830671 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:26:33 crc kubenswrapper[4837]: E1001 09:26:33.831940 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:26:48 crc kubenswrapper[4837]: I1001 09:26:48.817560 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:26:48 crc kubenswrapper[4837]: E1001 09:26:48.819318 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:27:00 crc kubenswrapper[4837]: I1001 09:27:00.816408 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:27:00 crc kubenswrapper[4837]: E1001 09:27:00.817263 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:27:11 crc kubenswrapper[4837]: I1001 09:27:11.817280 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:27:11 crc kubenswrapper[4837]: E1001 09:27:11.818545 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:27:22 crc kubenswrapper[4837]: I1001 09:27:22.816584 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:27:22 crc kubenswrapper[4837]: E1001 09:27:22.817584 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:27:33 crc kubenswrapper[4837]: I1001 09:27:33.822172 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:27:33 crc kubenswrapper[4837]: E1001 09:27:33.823411 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:27:48 crc kubenswrapper[4837]: I1001 09:27:48.815685 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:27:48 crc kubenswrapper[4837]: E1001 09:27:48.816793 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:01 crc kubenswrapper[4837]: I1001 09:28:01.816539 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:28:01 crc kubenswrapper[4837]: E1001 09:28:01.817305 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:12 crc kubenswrapper[4837]: I1001 09:28:12.816384 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:28:12 crc kubenswrapper[4837]: E1001 09:28:12.817348 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:15 crc kubenswrapper[4837]: I1001 09:28:15.351579 4837 scope.go:117] "RemoveContainer" containerID="e06df847d18f96f6dfa79def67a92183be9c94d8ef96aeaa6145792e50041898" Oct 01 09:28:15 crc kubenswrapper[4837]: I1001 09:28:15.372805 4837 scope.go:117] "RemoveContainer" containerID="56e65d1eff4a3636036e5920f7edfe7e6121f27821ccb5c9b828b491d74ff0b6" Oct 01 09:28:15 crc kubenswrapper[4837]: I1001 09:28:15.429420 4837 scope.go:117] "RemoveContainer" containerID="2f5dce9929f2b7c5ad55f0736e523c0a87988b7bef01ec619f5fab19f74fe05d" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.012533 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.015493 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.047583 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.125626 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.125725 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.125790 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrrfc\" (UniqueName: \"kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.228492 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.228550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrrfc\" (UniqueName: \"kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.228675 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.229197 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.229310 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.252210 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrrfc\" (UniqueName: \"kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc\") pod \"certified-operators-lvjzx\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.352606 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:17 crc kubenswrapper[4837]: I1001 09:28:17.799836 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:18 crc kubenswrapper[4837]: I1001 09:28:18.379612 4837 generic.go:334] "Generic (PLEG): container finished" podID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerID="19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7" exitCode=0 Oct 01 09:28:18 crc kubenswrapper[4837]: I1001 09:28:18.379674 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerDied","Data":"19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7"} Oct 01 09:28:18 crc kubenswrapper[4837]: I1001 09:28:18.379786 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerStarted","Data":"cb504c720234a519b09229a3bf73ef64fc6731b258eaa278a29d38e5c7afe36e"} Oct 01 09:28:19 crc kubenswrapper[4837]: I1001 09:28:19.393831 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerStarted","Data":"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941"} Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.569370 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.571675 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.591480 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.700453 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2pq\" (UniqueName: \"kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.700984 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.701202 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.802798 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.802875 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.802967 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2pq\" (UniqueName: \"kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.803510 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.803570 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.833897 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2pq\" (UniqueName: \"kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq\") pod \"redhat-marketplace-4sprq\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:20 crc kubenswrapper[4837]: I1001 09:28:20.896372 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:21 crc kubenswrapper[4837]: I1001 09:28:21.424202 4837 generic.go:334] "Generic (PLEG): container finished" podID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerID="641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941" exitCode=0 Oct 01 09:28:21 crc kubenswrapper[4837]: I1001 09:28:21.424287 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerDied","Data":"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941"} Oct 01 09:28:21 crc kubenswrapper[4837]: I1001 09:28:21.435729 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:22 crc kubenswrapper[4837]: I1001 09:28:22.440936 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerStarted","Data":"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba"} Oct 01 09:28:22 crc kubenswrapper[4837]: I1001 09:28:22.446645 4837 generic.go:334] "Generic (PLEG): container finished" podID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerID="bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425" exitCode=0 Oct 01 09:28:22 crc kubenswrapper[4837]: I1001 09:28:22.446728 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerDied","Data":"bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425"} Oct 01 09:28:22 crc kubenswrapper[4837]: I1001 09:28:22.446769 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerStarted","Data":"e9e8bc14d82645f29375e27c3b5fa6a99be9312d40e8b353309561493ef05143"} Oct 01 09:28:22 crc kubenswrapper[4837]: I1001 09:28:22.479119 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lvjzx" podStartSLOduration=2.751309172 podStartE2EDuration="6.479097059s" podCreationTimestamp="2025-10-01 09:28:16 +0000 UTC" firstStartedPulling="2025-10-01 09:28:18.381480991 +0000 UTC m=+8555.223088496" lastFinishedPulling="2025-10-01 09:28:22.109268898 +0000 UTC m=+8558.950876383" observedRunningTime="2025-10-01 09:28:22.475164943 +0000 UTC m=+8559.316772438" watchObservedRunningTime="2025-10-01 09:28:22.479097059 +0000 UTC m=+8559.320704514" Oct 01 09:28:23 crc kubenswrapper[4837]: I1001 09:28:23.456981 4837 generic.go:334] "Generic (PLEG): container finished" podID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerID="74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe" exitCode=0 Oct 01 09:28:23 crc kubenswrapper[4837]: I1001 09:28:23.457038 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerDied","Data":"74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe"} Oct 01 09:28:24 crc kubenswrapper[4837]: I1001 09:28:24.477414 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerStarted","Data":"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678"} Oct 01 09:28:24 crc kubenswrapper[4837]: I1001 09:28:24.526946 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4sprq" podStartSLOduration=3.115893752 podStartE2EDuration="4.526919899s" podCreationTimestamp="2025-10-01 09:28:20 +0000 UTC" firstStartedPulling="2025-10-01 09:28:22.44865112 +0000 UTC m=+8559.290258605" lastFinishedPulling="2025-10-01 09:28:23.859677297 +0000 UTC m=+8560.701284752" observedRunningTime="2025-10-01 09:28:24.513638972 +0000 UTC m=+8561.355246507" watchObservedRunningTime="2025-10-01 09:28:24.526919899 +0000 UTC m=+8561.368527354" Oct 01 09:28:24 crc kubenswrapper[4837]: I1001 09:28:24.815760 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:28:24 crc kubenswrapper[4837]: E1001 09:28:24.816021 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:27 crc kubenswrapper[4837]: I1001 09:28:27.353518 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:27 crc kubenswrapper[4837]: I1001 09:28:27.353892 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:27 crc kubenswrapper[4837]: I1001 09:28:27.404376 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:27 crc kubenswrapper[4837]: I1001 09:28:27.571969 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:28 crc kubenswrapper[4837]: I1001 09:28:28.566948 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:29 crc kubenswrapper[4837]: I1001 09:28:29.526154 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lvjzx" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="registry-server" containerID="cri-o://06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba" gracePeriod=2 Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.027084 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.122848 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content\") pod \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.123000 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrrfc\" (UniqueName: \"kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc\") pod \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.123054 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities\") pod \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\" (UID: \"9c54be9d-92aa-4472-83dd-ef722b43e5a2\") " Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.124150 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities" (OuterVolumeSpecName: "utilities") pod "9c54be9d-92aa-4472-83dd-ef722b43e5a2" (UID: "9c54be9d-92aa-4472-83dd-ef722b43e5a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.128264 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc" (OuterVolumeSpecName: "kube-api-access-rrrfc") pod "9c54be9d-92aa-4472-83dd-ef722b43e5a2" (UID: "9c54be9d-92aa-4472-83dd-ef722b43e5a2"). InnerVolumeSpecName "kube-api-access-rrrfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.165661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c54be9d-92aa-4472-83dd-ef722b43e5a2" (UID: "9c54be9d-92aa-4472-83dd-ef722b43e5a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.225947 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.225986 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrrfc\" (UniqueName: \"kubernetes.io/projected/9c54be9d-92aa-4472-83dd-ef722b43e5a2-kube-api-access-rrrfc\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.225999 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c54be9d-92aa-4472-83dd-ef722b43e5a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.543202 4837 generic.go:334] "Generic (PLEG): container finished" podID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerID="06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba" exitCode=0 Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.543275 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerDied","Data":"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba"} Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.543548 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvjzx" event={"ID":"9c54be9d-92aa-4472-83dd-ef722b43e5a2","Type":"ContainerDied","Data":"cb504c720234a519b09229a3bf73ef64fc6731b258eaa278a29d38e5c7afe36e"} Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.543575 4837 scope.go:117] "RemoveContainer" containerID="06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.543348 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvjzx" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.590574 4837 scope.go:117] "RemoveContainer" containerID="641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.599127 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.605450 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lvjzx"] Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.625727 4837 scope.go:117] "RemoveContainer" containerID="19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.667213 4837 scope.go:117] "RemoveContainer" containerID="06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba" Oct 01 09:28:30 crc kubenswrapper[4837]: E1001 09:28:30.667755 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba\": container with ID starting with 06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba not found: ID does not exist" containerID="06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.667784 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba"} err="failed to get container status \"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba\": rpc error: code = NotFound desc = could not find container \"06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba\": container with ID starting with 06b1a28c9d437bc847128e5a6137fafa86fda88ed9d4ed1f418c19d62781bbba not found: ID does not exist" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.667817 4837 scope.go:117] "RemoveContainer" containerID="641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941" Oct 01 09:28:30 crc kubenswrapper[4837]: E1001 09:28:30.668176 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941\": container with ID starting with 641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941 not found: ID does not exist" containerID="641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.668236 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941"} err="failed to get container status \"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941\": rpc error: code = NotFound desc = could not find container \"641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941\": container with ID starting with 641d999c393bb10428156ffbbd7f38b2d16c11bdd7f7abd7e4e79d3567e65941 not found: ID does not exist" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.668273 4837 scope.go:117] "RemoveContainer" containerID="19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7" Oct 01 09:28:30 crc kubenswrapper[4837]: E1001 09:28:30.668712 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7\": container with ID starting with 19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7 not found: ID does not exist" containerID="19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.668740 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7"} err="failed to get container status \"19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7\": rpc error: code = NotFound desc = could not find container \"19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7\": container with ID starting with 19e9d985c955b550f962f9ef3455f0c4f248bf178184bd6853b72999bbc962e7 not found: ID does not exist" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.896874 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.897056 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:30 crc kubenswrapper[4837]: I1001 09:28:30.964601 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:31 crc kubenswrapper[4837]: I1001 09:28:31.627471 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:31 crc kubenswrapper[4837]: I1001 09:28:31.829844 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" path="/var/lib/kubelet/pods/9c54be9d-92aa-4472-83dd-ef722b43e5a2/volumes" Oct 01 09:28:32 crc kubenswrapper[4837]: I1001 09:28:32.972040 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:34 crc kubenswrapper[4837]: I1001 09:28:34.599241 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4sprq" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="registry-server" containerID="cri-o://3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678" gracePeriod=2 Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.116778 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.233594 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content\") pod \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.233632 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities\") pod \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.233760 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk2pq\" (UniqueName: \"kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq\") pod \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\" (UID: \"287e0cba-4e0f-45d2-82b1-ae0ed444d905\") " Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.235932 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities" (OuterVolumeSpecName: "utilities") pod "287e0cba-4e0f-45d2-82b1-ae0ed444d905" (UID: "287e0cba-4e0f-45d2-82b1-ae0ed444d905"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.239956 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq" (OuterVolumeSpecName: "kube-api-access-qk2pq") pod "287e0cba-4e0f-45d2-82b1-ae0ed444d905" (UID: "287e0cba-4e0f-45d2-82b1-ae0ed444d905"). InnerVolumeSpecName "kube-api-access-qk2pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.249537 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "287e0cba-4e0f-45d2-82b1-ae0ed444d905" (UID: "287e0cba-4e0f-45d2-82b1-ae0ed444d905"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.335898 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.335938 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/287e0cba-4e0f-45d2-82b1-ae0ed444d905-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.335949 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk2pq\" (UniqueName: \"kubernetes.io/projected/287e0cba-4e0f-45d2-82b1-ae0ed444d905-kube-api-access-qk2pq\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.611212 4837 generic.go:334] "Generic (PLEG): container finished" podID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerID="3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678" exitCode=0 Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.611279 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4sprq" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.611293 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerDied","Data":"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678"} Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.613201 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4sprq" event={"ID":"287e0cba-4e0f-45d2-82b1-ae0ed444d905","Type":"ContainerDied","Data":"e9e8bc14d82645f29375e27c3b5fa6a99be9312d40e8b353309561493ef05143"} Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.613317 4837 scope.go:117] "RemoveContainer" containerID="3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.649582 4837 scope.go:117] "RemoveContainer" containerID="74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.659576 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.669471 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4sprq"] Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.669943 4837 scope.go:117] "RemoveContainer" containerID="bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.724040 4837 scope.go:117] "RemoveContainer" containerID="3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678" Oct 01 09:28:35 crc kubenswrapper[4837]: E1001 09:28:35.724426 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678\": container with ID starting with 3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678 not found: ID does not exist" containerID="3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.724464 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678"} err="failed to get container status \"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678\": rpc error: code = NotFound desc = could not find container \"3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678\": container with ID starting with 3b36a2527cc86d20efd3507b2112f763a5f9e24ce97bc07ec866fa166ebcb678 not found: ID does not exist" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.724484 4837 scope.go:117] "RemoveContainer" containerID="74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe" Oct 01 09:28:35 crc kubenswrapper[4837]: E1001 09:28:35.724906 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe\": container with ID starting with 74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe not found: ID does not exist" containerID="74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.725032 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe"} err="failed to get container status \"74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe\": rpc error: code = NotFound desc = could not find container \"74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe\": container with ID starting with 74c5617ec58d8012d1c3a07e3ae53e338661c1beff8b0add1835fdccade5f7fe not found: ID does not exist" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.725136 4837 scope.go:117] "RemoveContainer" containerID="bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425" Oct 01 09:28:35 crc kubenswrapper[4837]: E1001 09:28:35.725739 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425\": container with ID starting with bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425 not found: ID does not exist" containerID="bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.725792 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425"} err="failed to get container status \"bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425\": rpc error: code = NotFound desc = could not find container \"bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425\": container with ID starting with bc83d44cb1b9f9d9ca279e4890aa845257ab47472b2f6842c616eced615f4425 not found: ID does not exist" Oct 01 09:28:35 crc kubenswrapper[4837]: I1001 09:28:35.830188 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" path="/var/lib/kubelet/pods/287e0cba-4e0f-45d2-82b1-ae0ed444d905/volumes" Oct 01 09:28:39 crc kubenswrapper[4837]: I1001 09:28:39.816965 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:28:39 crc kubenswrapper[4837]: E1001 09:28:39.817966 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.979468 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982526 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982555 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982584 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982594 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982619 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="extract-utilities" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982629 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="extract-utilities" Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982664 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="extract-utilities" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982673 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="extract-utilities" Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982711 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="extract-content" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982720 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="extract-content" Oct 01 09:28:40 crc kubenswrapper[4837]: E1001 09:28:40.982739 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="extract-content" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.982763 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="extract-content" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.983007 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="287e0cba-4e0f-45d2-82b1-ae0ed444d905" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.983047 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c54be9d-92aa-4472-83dd-ef722b43e5a2" containerName="registry-server" Oct 01 09:28:40 crc kubenswrapper[4837]: I1001 09:28:40.984854 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.013481 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.064674 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7bnc\" (UniqueName: \"kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.064755 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.064816 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.166256 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7bnc\" (UniqueName: \"kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.166316 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.166363 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.166953 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.166949 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.186064 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7bnc\" (UniqueName: \"kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc\") pod \"redhat-operators-rc284\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.307472 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:41 crc kubenswrapper[4837]: I1001 09:28:41.779516 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:28:42 crc kubenswrapper[4837]: I1001 09:28:42.681937 4837 generic.go:334] "Generic (PLEG): container finished" podID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerID="08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf" exitCode=0 Oct 01 09:28:42 crc kubenswrapper[4837]: I1001 09:28:42.682063 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerDied","Data":"08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf"} Oct 01 09:28:42 crc kubenswrapper[4837]: I1001 09:28:42.682394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerStarted","Data":"9b92aceadf75c41ea7eb5d113f2f4568af4aa989ee8e0403aa13b611576d1331"} Oct 01 09:28:42 crc kubenswrapper[4837]: I1001 09:28:42.685680 4837 generic.go:334] "Generic (PLEG): container finished" podID="7b2e57be-af14-4e5b-8751-dc110da24b25" containerID="e31bc9d0a521d4afe5881cfa01f9506b8158365e64a2751af1af9330c97b3a76" exitCode=0 Oct 01 09:28:42 crc kubenswrapper[4837]: I1001 09:28:42.685744 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" event={"ID":"7b2e57be-af14-4e5b-8751-dc110da24b25","Type":"ContainerDied","Data":"e31bc9d0a521d4afe5881cfa01f9506b8158365e64a2751af1af9330c97b3a76"} Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.179302 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.329404 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle\") pod \"7b2e57be-af14-4e5b-8751-dc110da24b25\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.329640 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key\") pod \"7b2e57be-af14-4e5b-8751-dc110da24b25\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.329733 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory\") pod \"7b2e57be-af14-4e5b-8751-dc110da24b25\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.329805 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lgnv\" (UniqueName: \"kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv\") pod \"7b2e57be-af14-4e5b-8751-dc110da24b25\" (UID: \"7b2e57be-af14-4e5b-8751-dc110da24b25\") " Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.337327 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7b2e57be-af14-4e5b-8751-dc110da24b25" (UID: "7b2e57be-af14-4e5b-8751-dc110da24b25"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.337427 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv" (OuterVolumeSpecName: "kube-api-access-4lgnv") pod "7b2e57be-af14-4e5b-8751-dc110da24b25" (UID: "7b2e57be-af14-4e5b-8751-dc110da24b25"). InnerVolumeSpecName "kube-api-access-4lgnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.366887 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b2e57be-af14-4e5b-8751-dc110da24b25" (UID: "7b2e57be-af14-4e5b-8751-dc110da24b25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.381026 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory" (OuterVolumeSpecName: "inventory") pod "7b2e57be-af14-4e5b-8751-dc110da24b25" (UID: "7b2e57be-af14-4e5b-8751-dc110da24b25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.432760 4837 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.432791 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.432803 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b2e57be-af14-4e5b-8751-dc110da24b25-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.432815 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lgnv\" (UniqueName: \"kubernetes.io/projected/7b2e57be-af14-4e5b-8751-dc110da24b25-kube-api-access-4lgnv\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.705541 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" event={"ID":"7b2e57be-af14-4e5b-8751-dc110da24b25","Type":"ContainerDied","Data":"f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0"} Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.705582 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f18d2574e217c030c8b3341a89a0fab88713d8678f2781aca9f41c3bd9762ed0" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.705601 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-trbtg" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.708548 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerStarted","Data":"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448"} Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.820282 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-networker-wqxpl"] Oct 01 09:28:44 crc kubenswrapper[4837]: E1001 09:28:44.821057 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2e57be-af14-4e5b-8751-dc110da24b25" containerName="bootstrap-openstack-openstack-networker" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.821078 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2e57be-af14-4e5b-8751-dc110da24b25" containerName="bootstrap-openstack-openstack-networker" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.821317 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2e57be-af14-4e5b-8751-dc110da24b25" containerName="bootstrap-openstack-openstack-networker" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.822206 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.823904 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.827479 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.834999 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-wqxpl"] Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.945897 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.946201 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgw2\" (UniqueName: \"kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:44 crc kubenswrapper[4837]: I1001 09:28:44.946583 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.049032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgw2\" (UniqueName: \"kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.049228 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.049437 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.055891 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.063417 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.080969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgw2\" (UniqueName: \"kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2\") pod \"download-cache-openstack-openstack-networker-wqxpl\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.143202 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:28:45 crc kubenswrapper[4837]: W1001 09:28:45.742166 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf90a4c39_8e54_4d88_8e72_b8ad8caadd2f.slice/crio-d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940 WatchSource:0}: Error finding container d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940: Status 404 returned error can't find the container with id d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940 Oct 01 09:28:45 crc kubenswrapper[4837]: I1001 09:28:45.755071 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-wqxpl"] Oct 01 09:28:46 crc kubenswrapper[4837]: I1001 09:28:46.750997 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" event={"ID":"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f","Type":"ContainerStarted","Data":"d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940"} Oct 01 09:28:47 crc kubenswrapper[4837]: I1001 09:28:47.762828 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" event={"ID":"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f","Type":"ContainerStarted","Data":"b82cc028d362c6fd2cf03a0b2f2f5bf54363f788bb4081ef0b9026393440d4f0"} Oct 01 09:28:47 crc kubenswrapper[4837]: I1001 09:28:47.768273 4837 generic.go:334] "Generic (PLEG): container finished" podID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerID="f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448" exitCode=0 Oct 01 09:28:47 crc kubenswrapper[4837]: I1001 09:28:47.768354 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerDied","Data":"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448"} Oct 01 09:28:47 crc kubenswrapper[4837]: I1001 09:28:47.779948 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" podStartSLOduration=3.256374462 podStartE2EDuration="3.779924237s" podCreationTimestamp="2025-10-01 09:28:44 +0000 UTC" firstStartedPulling="2025-10-01 09:28:45.746227644 +0000 UTC m=+8582.587835099" lastFinishedPulling="2025-10-01 09:28:46.269777409 +0000 UTC m=+8583.111384874" observedRunningTime="2025-10-01 09:28:47.778985323 +0000 UTC m=+8584.620592798" watchObservedRunningTime="2025-10-01 09:28:47.779924237 +0000 UTC m=+8584.621531732" Oct 01 09:28:48 crc kubenswrapper[4837]: I1001 09:28:48.779128 4837 generic.go:334] "Generic (PLEG): container finished" podID="18443fd2-82a7-4be3-9ad7-8be4777ec6e5" containerID="fd291b579eb3ff0e0ac865fb871de393dfdf22480ec9a5667ea4b7ad9659d9b4" exitCode=0 Oct 01 09:28:48 crc kubenswrapper[4837]: I1001 09:28:48.779190 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" event={"ID":"18443fd2-82a7-4be3-9ad7-8be4777ec6e5","Type":"ContainerDied","Data":"fd291b579eb3ff0e0ac865fb871de393dfdf22480ec9a5667ea4b7ad9659d9b4"} Oct 01 09:28:48 crc kubenswrapper[4837]: I1001 09:28:48.782464 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerStarted","Data":"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0"} Oct 01 09:28:48 crc kubenswrapper[4837]: I1001 09:28:48.819866 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rc284" podStartSLOduration=3.218987877 podStartE2EDuration="8.819851661s" podCreationTimestamp="2025-10-01 09:28:40 +0000 UTC" firstStartedPulling="2025-10-01 09:28:42.684672676 +0000 UTC m=+8579.526280131" lastFinishedPulling="2025-10-01 09:28:48.28553646 +0000 UTC m=+8585.127143915" observedRunningTime="2025-10-01 09:28:48.818503608 +0000 UTC m=+8585.660111063" watchObservedRunningTime="2025-10-01 09:28:48.819851661 +0000 UTC m=+8585.661459106" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.255636 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.357966 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle\") pod \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.358125 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key\") pod \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.358147 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7phr7\" (UniqueName: \"kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7\") pod \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.358201 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory\") pod \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\" (UID: \"18443fd2-82a7-4be3-9ad7-8be4777ec6e5\") " Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.365180 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7" (OuterVolumeSpecName: "kube-api-access-7phr7") pod "18443fd2-82a7-4be3-9ad7-8be4777ec6e5" (UID: "18443fd2-82a7-4be3-9ad7-8be4777ec6e5"). InnerVolumeSpecName "kube-api-access-7phr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.367543 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "18443fd2-82a7-4be3-9ad7-8be4777ec6e5" (UID: "18443fd2-82a7-4be3-9ad7-8be4777ec6e5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.400220 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory" (OuterVolumeSpecName: "inventory") pod "18443fd2-82a7-4be3-9ad7-8be4777ec6e5" (UID: "18443fd2-82a7-4be3-9ad7-8be4777ec6e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.422278 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18443fd2-82a7-4be3-9ad7-8be4777ec6e5" (UID: "18443fd2-82a7-4be3-9ad7-8be4777ec6e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.460264 4837 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.460308 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.460320 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7phr7\" (UniqueName: \"kubernetes.io/projected/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-kube-api-access-7phr7\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.460330 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18443fd2-82a7-4be3-9ad7-8be4777ec6e5-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.805167 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" event={"ID":"18443fd2-82a7-4be3-9ad7-8be4777ec6e5","Type":"ContainerDied","Data":"6b8568d8c3dcf3c2b3d2ebfb2857851b478066c2371af12bec6591cc2f58df84"} Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.805232 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b8568d8c3dcf3c2b3d2ebfb2857851b478066c2371af12bec6591cc2f58df84" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.805268 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-lhpww" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.884899 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nbbwp"] Oct 01 09:28:50 crc kubenswrapper[4837]: E1001 09:28:50.885350 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18443fd2-82a7-4be3-9ad7-8be4777ec6e5" containerName="bootstrap-openstack-openstack-cell1" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.885370 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="18443fd2-82a7-4be3-9ad7-8be4777ec6e5" containerName="bootstrap-openstack-openstack-cell1" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.885637 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="18443fd2-82a7-4be3-9ad7-8be4777ec6e5" containerName="bootstrap-openstack-openstack-cell1" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.886668 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.889023 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.889787 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.898363 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nbbwp"] Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.980194 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.980544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbcrb\" (UniqueName: \"kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:50 crc kubenswrapper[4837]: I1001 09:28:50.980782 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.082278 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.082743 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbcrb\" (UniqueName: \"kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.082781 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.087038 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.087770 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.101388 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbcrb\" (UniqueName: \"kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb\") pod \"download-cache-openstack-openstack-cell1-nbbwp\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.209226 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.308329 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.310916 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.738841 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nbbwp"] Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.813334 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" event={"ID":"4a18e58a-b31f-43b9-8fac-5d55c9e65621","Type":"ContainerStarted","Data":"b3f598ce6cf6230d1e8d15a910e50c2c06ac2c50b4e3f5442b2c072404b82f66"} Oct 01 09:28:51 crc kubenswrapper[4837]: I1001 09:28:51.817305 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:28:51 crc kubenswrapper[4837]: E1001 09:28:51.817601 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:28:52 crc kubenswrapper[4837]: I1001 09:28:52.376964 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rc284" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" probeResult="failure" output=< Oct 01 09:28:52 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:28:52 crc kubenswrapper[4837]: > Oct 01 09:28:52 crc kubenswrapper[4837]: I1001 09:28:52.825337 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" event={"ID":"4a18e58a-b31f-43b9-8fac-5d55c9e65621","Type":"ContainerStarted","Data":"d52aad826d232cd51ef8d6f1a08894d45a9fc6ddc02835281a392952208017e5"} Oct 01 09:28:52 crc kubenswrapper[4837]: I1001 09:28:52.858300 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" podStartSLOduration=2.373077661 podStartE2EDuration="2.858266762s" podCreationTimestamp="2025-10-01 09:28:50 +0000 UTC" firstStartedPulling="2025-10-01 09:28:51.746268014 +0000 UTC m=+8588.587875479" lastFinishedPulling="2025-10-01 09:28:52.231457065 +0000 UTC m=+8589.073064580" observedRunningTime="2025-10-01 09:28:52.840199697 +0000 UTC m=+8589.681807182" watchObservedRunningTime="2025-10-01 09:28:52.858266762 +0000 UTC m=+8589.699874257" Oct 01 09:29:02 crc kubenswrapper[4837]: I1001 09:29:02.356739 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rc284" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" probeResult="failure" output=< Oct 01 09:29:02 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:29:02 crc kubenswrapper[4837]: > Oct 01 09:29:03 crc kubenswrapper[4837]: I1001 09:29:03.822218 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:29:03 crc kubenswrapper[4837]: E1001 09:29:03.822761 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:29:11 crc kubenswrapper[4837]: I1001 09:29:11.389748 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:29:11 crc kubenswrapper[4837]: I1001 09:29:11.477502 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:29:11 crc kubenswrapper[4837]: I1001 09:29:11.644472 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.051994 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rc284" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" containerID="cri-o://00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0" gracePeriod=2 Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.623993 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.788219 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7bnc\" (UniqueName: \"kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc\") pod \"64c53279-2bbd-40e4-a487-c049d6d09abc\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.788312 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content\") pod \"64c53279-2bbd-40e4-a487-c049d6d09abc\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.788463 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities\") pod \"64c53279-2bbd-40e4-a487-c049d6d09abc\" (UID: \"64c53279-2bbd-40e4-a487-c049d6d09abc\") " Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.789104 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities" (OuterVolumeSpecName: "utilities") pod "64c53279-2bbd-40e4-a487-c049d6d09abc" (UID: "64c53279-2bbd-40e4-a487-c049d6d09abc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.793380 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc" (OuterVolumeSpecName: "kube-api-access-r7bnc") pod "64c53279-2bbd-40e4-a487-c049d6d09abc" (UID: "64c53279-2bbd-40e4-a487-c049d6d09abc"). InnerVolumeSpecName "kube-api-access-r7bnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.879472 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64c53279-2bbd-40e4-a487-c049d6d09abc" (UID: "64c53279-2bbd-40e4-a487-c049d6d09abc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.890549 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.890584 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7bnc\" (UniqueName: \"kubernetes.io/projected/64c53279-2bbd-40e4-a487-c049d6d09abc-kube-api-access-r7bnc\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:13 crc kubenswrapper[4837]: I1001 09:29:13.890597 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64c53279-2bbd-40e4-a487-c049d6d09abc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.071968 4837 generic.go:334] "Generic (PLEG): container finished" podID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerID="00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0" exitCode=0 Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.072010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerDied","Data":"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0"} Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.072036 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rc284" event={"ID":"64c53279-2bbd-40e4-a487-c049d6d09abc","Type":"ContainerDied","Data":"9b92aceadf75c41ea7eb5d113f2f4568af4aa989ee8e0403aa13b611576d1331"} Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.072052 4837 scope.go:117] "RemoveContainer" containerID="00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.072215 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rc284" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.116348 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.120922 4837 scope.go:117] "RemoveContainer" containerID="f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.127414 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rc284"] Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.147263 4837 scope.go:117] "RemoveContainer" containerID="08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.242553 4837 scope.go:117] "RemoveContainer" containerID="00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0" Oct 01 09:29:14 crc kubenswrapper[4837]: E1001 09:29:14.243116 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0\": container with ID starting with 00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0 not found: ID does not exist" containerID="00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.243186 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0"} err="failed to get container status \"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0\": rpc error: code = NotFound desc = could not find container \"00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0\": container with ID starting with 00cf48239c3680e796d00949f5682765d50ac7b50dd0aa6ebc36bafba36333d0 not found: ID does not exist" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.243252 4837 scope.go:117] "RemoveContainer" containerID="f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448" Oct 01 09:29:14 crc kubenswrapper[4837]: E1001 09:29:14.243993 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448\": container with ID starting with f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448 not found: ID does not exist" containerID="f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.244051 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448"} err="failed to get container status \"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448\": rpc error: code = NotFound desc = could not find container \"f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448\": container with ID starting with f5596bd77ada0f800d371702784efc50c94b5c1673c2f0a2cf483f3317871448 not found: ID does not exist" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.244095 4837 scope.go:117] "RemoveContainer" containerID="08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf" Oct 01 09:29:14 crc kubenswrapper[4837]: E1001 09:29:14.244659 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf\": container with ID starting with 08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf not found: ID does not exist" containerID="08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf" Oct 01 09:29:14 crc kubenswrapper[4837]: I1001 09:29:14.244747 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf"} err="failed to get container status \"08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf\": rpc error: code = NotFound desc = could not find container \"08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf\": container with ID starting with 08707c46f03e566bb322609ac22f366af49996f3edfc7e4bb2ad07a72776b3cf not found: ID does not exist" Oct 01 09:29:15 crc kubenswrapper[4837]: I1001 09:29:15.816640 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:29:15 crc kubenswrapper[4837]: E1001 09:29:15.817728 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:29:15 crc kubenswrapper[4837]: I1001 09:29:15.833316 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" path="/var/lib/kubelet/pods/64c53279-2bbd-40e4-a487-c049d6d09abc/volumes" Oct 01 09:29:29 crc kubenswrapper[4837]: I1001 09:29:29.816415 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:29:29 crc kubenswrapper[4837]: E1001 09:29:29.817131 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:29:40 crc kubenswrapper[4837]: I1001 09:29:40.817046 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:29:40 crc kubenswrapper[4837]: E1001 09:29:40.818290 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:29:55 crc kubenswrapper[4837]: I1001 09:29:55.554871 4837 generic.go:334] "Generic (PLEG): container finished" podID="f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" containerID="b82cc028d362c6fd2cf03a0b2f2f5bf54363f788bb4081ef0b9026393440d4f0" exitCode=0 Oct 01 09:29:55 crc kubenswrapper[4837]: I1001 09:29:55.554951 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" event={"ID":"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f","Type":"ContainerDied","Data":"b82cc028d362c6fd2cf03a0b2f2f5bf54363f788bb4081ef0b9026393440d4f0"} Oct 01 09:29:55 crc kubenswrapper[4837]: I1001 09:29:55.816315 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:29:55 crc kubenswrapper[4837]: E1001 09:29:55.816675 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.008995 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.191262 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key\") pod \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.192107 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory\") pod \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.192321 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tgw2\" (UniqueName: \"kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2\") pod \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\" (UID: \"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f\") " Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.198971 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2" (OuterVolumeSpecName: "kube-api-access-6tgw2") pod "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" (UID: "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f"). InnerVolumeSpecName "kube-api-access-6tgw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.223101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" (UID: "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.242548 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory" (OuterVolumeSpecName: "inventory") pod "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" (UID: "f90a4c39-8e54-4d88-8e72-b8ad8caadd2f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.294747 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.294774 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.294784 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tgw2\" (UniqueName: \"kubernetes.io/projected/f90a4c39-8e54-4d88-8e72-b8ad8caadd2f-kube-api-access-6tgw2\") on node \"crc\" DevicePath \"\"" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.604681 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" event={"ID":"f90a4c39-8e54-4d88-8e72-b8ad8caadd2f","Type":"ContainerDied","Data":"d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940"} Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.604771 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9d0cdca4ce08a2b9bb4ad70508a99be9d91bb0385348b9ba5d772dc11939940" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.604812 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-wqxpl" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701261 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-networker-547hg"] Oct 01 09:29:57 crc kubenswrapper[4837]: E1001 09:29:57.701676 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701689 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" Oct 01 09:29:57 crc kubenswrapper[4837]: E1001 09:29:57.701718 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="extract-content" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701724 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="extract-content" Oct 01 09:29:57 crc kubenswrapper[4837]: E1001 09:29:57.701755 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" containerName="download-cache-openstack-openstack-networker" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701765 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" containerName="download-cache-openstack-openstack-networker" Oct 01 09:29:57 crc kubenswrapper[4837]: E1001 09:29:57.701783 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="extract-utilities" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701790 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="extract-utilities" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.701981 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c53279-2bbd-40e4-a487-c049d6d09abc" containerName="registry-server" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.702002 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90a4c39-8e54-4d88-8e72-b8ad8caadd2f" containerName="download-cache-openstack-openstack-networker" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.702726 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.705042 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.705461 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.714228 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-547hg"] Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.804751 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm4d2\" (UniqueName: \"kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.804921 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.805027 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.906473 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.906526 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.906659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm4d2\" (UniqueName: \"kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.911983 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.912359 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:57 crc kubenswrapper[4837]: I1001 09:29:57.922789 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm4d2\" (UniqueName: \"kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2\") pod \"configure-network-openstack-openstack-networker-547hg\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:58 crc kubenswrapper[4837]: I1001 09:29:58.021452 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:29:58 crc kubenswrapper[4837]: I1001 09:29:58.525988 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-547hg"] Oct 01 09:29:58 crc kubenswrapper[4837]: W1001 09:29:58.537199 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb189babd_cae5_47d1_bc90_c8a49033576e.slice/crio-4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b WatchSource:0}: Error finding container 4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b: Status 404 returned error can't find the container with id 4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b Oct 01 09:29:58 crc kubenswrapper[4837]: I1001 09:29:58.616718 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-547hg" event={"ID":"b189babd-cae5-47d1-bc90-c8a49033576e","Type":"ContainerStarted","Data":"4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b"} Oct 01 09:29:59 crc kubenswrapper[4837]: I1001 09:29:59.630743 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-547hg" event={"ID":"b189babd-cae5-47d1-bc90-c8a49033576e","Type":"ContainerStarted","Data":"2015a8484a569162f636303dcaba2f27d418262cb8df9e558c40966750615cbe"} Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.156734 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-networker-547hg" podStartSLOduration=2.603450801 podStartE2EDuration="3.156708787s" podCreationTimestamp="2025-10-01 09:29:57 +0000 UTC" firstStartedPulling="2025-10-01 09:29:58.540312366 +0000 UTC m=+8655.381919821" lastFinishedPulling="2025-10-01 09:29:59.093570312 +0000 UTC m=+8655.935177807" observedRunningTime="2025-10-01 09:29:59.663932329 +0000 UTC m=+8656.505539784" watchObservedRunningTime="2025-10-01 09:30:00.156708787 +0000 UTC m=+8656.998316262" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.163147 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl"] Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.165193 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.169072 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.169285 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.179687 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl"] Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.260588 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.260733 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv8v5\" (UniqueName: \"kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.260822 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.363138 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.363273 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv8v5\" (UniqueName: \"kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.363360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.364829 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.374276 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.383488 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv8v5\" (UniqueName: \"kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5\") pod \"collect-profiles-29321850-drzrl\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.488249 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:00 crc kubenswrapper[4837]: I1001 09:30:00.986201 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl"] Oct 01 09:30:01 crc kubenswrapper[4837]: I1001 09:30:01.654616 4837 generic.go:334] "Generic (PLEG): container finished" podID="1f79b0eb-b8db-4fda-8573-329771458ba8" containerID="79a195d572e184aa02b0001aa13c210c0da0627cc5affeb645259c93dadd62ac" exitCode=0 Oct 01 09:30:01 crc kubenswrapper[4837]: I1001 09:30:01.654716 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" event={"ID":"1f79b0eb-b8db-4fda-8573-329771458ba8","Type":"ContainerDied","Data":"79a195d572e184aa02b0001aa13c210c0da0627cc5affeb645259c93dadd62ac"} Oct 01 09:30:01 crc kubenswrapper[4837]: I1001 09:30:01.655029 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" event={"ID":"1f79b0eb-b8db-4fda-8573-329771458ba8","Type":"ContainerStarted","Data":"cb08efce9d78b61e688d224315ce686ff39c89a4307678bcacc18e7b8fbebe91"} Oct 01 09:30:02 crc kubenswrapper[4837]: I1001 09:30:02.998201 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.121931 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume\") pod \"1f79b0eb-b8db-4fda-8573-329771458ba8\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.122110 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv8v5\" (UniqueName: \"kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5\") pod \"1f79b0eb-b8db-4fda-8573-329771458ba8\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.122240 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume\") pod \"1f79b0eb-b8db-4fda-8573-329771458ba8\" (UID: \"1f79b0eb-b8db-4fda-8573-329771458ba8\") " Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.123190 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume" (OuterVolumeSpecName: "config-volume") pod "1f79b0eb-b8db-4fda-8573-329771458ba8" (UID: "1f79b0eb-b8db-4fda-8573-329771458ba8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.129204 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5" (OuterVolumeSpecName: "kube-api-access-jv8v5") pod "1f79b0eb-b8db-4fda-8573-329771458ba8" (UID: "1f79b0eb-b8db-4fda-8573-329771458ba8"). InnerVolumeSpecName "kube-api-access-jv8v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.129401 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1f79b0eb-b8db-4fda-8573-329771458ba8" (UID: "1f79b0eb-b8db-4fda-8573-329771458ba8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.224944 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f79b0eb-b8db-4fda-8573-329771458ba8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.225008 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv8v5\" (UniqueName: \"kubernetes.io/projected/1f79b0eb-b8db-4fda-8573-329771458ba8-kube-api-access-jv8v5\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.225027 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f79b0eb-b8db-4fda-8573-329771458ba8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.680395 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" event={"ID":"1f79b0eb-b8db-4fda-8573-329771458ba8","Type":"ContainerDied","Data":"cb08efce9d78b61e688d224315ce686ff39c89a4307678bcacc18e7b8fbebe91"} Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.680765 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb08efce9d78b61e688d224315ce686ff39c89a4307678bcacc18e7b8fbebe91" Oct 01 09:30:03 crc kubenswrapper[4837]: I1001 09:30:03.680488 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl" Oct 01 09:30:04 crc kubenswrapper[4837]: I1001 09:30:04.099504 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66"] Oct 01 09:30:04 crc kubenswrapper[4837]: I1001 09:30:04.110933 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321805-mng66"] Oct 01 09:30:05 crc kubenswrapper[4837]: I1001 09:30:05.827271 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a3802d-4081-433c-a6e1-40f2f0c2f0b0" path="/var/lib/kubelet/pods/30a3802d-4081-433c-a6e1-40f2f0c2f0b0/volumes" Oct 01 09:30:10 crc kubenswrapper[4837]: I1001 09:30:10.816212 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:30:10 crc kubenswrapper[4837]: E1001 09:30:10.817269 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:30:15 crc kubenswrapper[4837]: I1001 09:30:15.616565 4837 scope.go:117] "RemoveContainer" containerID="58890a460f768b7ab805c5c239375dcab797cac9f839f99c4208dc616f49952e" Oct 01 09:30:20 crc kubenswrapper[4837]: I1001 09:30:20.866097 4837 generic.go:334] "Generic (PLEG): container finished" podID="4a18e58a-b31f-43b9-8fac-5d55c9e65621" containerID="d52aad826d232cd51ef8d6f1a08894d45a9fc6ddc02835281a392952208017e5" exitCode=0 Oct 01 09:30:20 crc kubenswrapper[4837]: I1001 09:30:20.866155 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" event={"ID":"4a18e58a-b31f-43b9-8fac-5d55c9e65621","Type":"ContainerDied","Data":"d52aad826d232cd51ef8d6f1a08894d45a9fc6ddc02835281a392952208017e5"} Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.473865 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.649494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbcrb\" (UniqueName: \"kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb\") pod \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.650279 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key\") pod \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.650387 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory\") pod \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\" (UID: \"4a18e58a-b31f-43b9-8fac-5d55c9e65621\") " Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.657130 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb" (OuterVolumeSpecName: "kube-api-access-zbcrb") pod "4a18e58a-b31f-43b9-8fac-5d55c9e65621" (UID: "4a18e58a-b31f-43b9-8fac-5d55c9e65621"). InnerVolumeSpecName "kube-api-access-zbcrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.689082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory" (OuterVolumeSpecName: "inventory") pod "4a18e58a-b31f-43b9-8fac-5d55c9e65621" (UID: "4a18e58a-b31f-43b9-8fac-5d55c9e65621"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.689163 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a18e58a-b31f-43b9-8fac-5d55c9e65621" (UID: "4a18e58a-b31f-43b9-8fac-5d55c9e65621"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.753812 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.753863 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a18e58a-b31f-43b9-8fac-5d55c9e65621-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.753886 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbcrb\" (UniqueName: \"kubernetes.io/projected/4a18e58a-b31f-43b9-8fac-5d55c9e65621-kube-api-access-zbcrb\") on node \"crc\" DevicePath \"\"" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.891241 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" event={"ID":"4a18e58a-b31f-43b9-8fac-5d55c9e65621","Type":"ContainerDied","Data":"b3f598ce6cf6230d1e8d15a910e50c2c06ac2c50b4e3f5442b2c072404b82f66"} Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.891300 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3f598ce6cf6230d1e8d15a910e50c2c06ac2c50b4e3f5442b2c072404b82f66" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.891339 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nbbwp" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.991575 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-hpq4d"] Oct 01 09:30:22 crc kubenswrapper[4837]: E1001 09:30:22.992036 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a18e58a-b31f-43b9-8fac-5d55c9e65621" containerName="download-cache-openstack-openstack-cell1" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.992054 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a18e58a-b31f-43b9-8fac-5d55c9e65621" containerName="download-cache-openstack-openstack-cell1" Oct 01 09:30:22 crc kubenswrapper[4837]: E1001 09:30:22.992089 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f79b0eb-b8db-4fda-8573-329771458ba8" containerName="collect-profiles" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.992095 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f79b0eb-b8db-4fda-8573-329771458ba8" containerName="collect-profiles" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.992274 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f79b0eb-b8db-4fda-8573-329771458ba8" containerName="collect-profiles" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.992291 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a18e58a-b31f-43b9-8fac-5d55c9e65621" containerName="download-cache-openstack-openstack-cell1" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.993004 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.995496 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:30:22 crc kubenswrapper[4837]: I1001 09:30:22.997044 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.008132 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-hpq4d"] Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.164298 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hvtv\" (UniqueName: \"kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.164392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.164438 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.266457 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.266684 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hvtv\" (UniqueName: \"kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.266861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.272589 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.272911 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.303323 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hvtv\" (UniqueName: \"kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv\") pod \"configure-network-openstack-openstack-cell1-hpq4d\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.339441 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:30:23 crc kubenswrapper[4837]: I1001 09:30:23.912366 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-hpq4d"] Oct 01 09:30:24 crc kubenswrapper[4837]: I1001 09:30:24.916074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" event={"ID":"9ccf0a84-b87b-4a54-b268-8b55d19dba49","Type":"ContainerStarted","Data":"b02a0dace5b2b9051452b9bbb0dcf71f05042e9ac99c54c79c701b53333d2784"} Oct 01 09:30:24 crc kubenswrapper[4837]: I1001 09:30:24.916410 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" event={"ID":"9ccf0a84-b87b-4a54-b268-8b55d19dba49","Type":"ContainerStarted","Data":"63257ddc5118bbf9f674d59c9cbeeb699846b87a6b57e31855320ceafa9066b5"} Oct 01 09:30:24 crc kubenswrapper[4837]: I1001 09:30:24.933845 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" podStartSLOduration=2.4366033910000002 podStartE2EDuration="2.933822478s" podCreationTimestamp="2025-10-01 09:30:22 +0000 UTC" firstStartedPulling="2025-10-01 09:30:23.922729683 +0000 UTC m=+8680.764337138" lastFinishedPulling="2025-10-01 09:30:24.41994873 +0000 UTC m=+8681.261556225" observedRunningTime="2025-10-01 09:30:24.930629289 +0000 UTC m=+8681.772236764" watchObservedRunningTime="2025-10-01 09:30:24.933822478 +0000 UTC m=+8681.775429943" Oct 01 09:30:25 crc kubenswrapper[4837]: I1001 09:30:25.815784 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:30:25 crc kubenswrapper[4837]: E1001 09:30:25.816279 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:30:36 crc kubenswrapper[4837]: I1001 09:30:36.816967 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:30:36 crc kubenswrapper[4837]: E1001 09:30:36.817969 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:30:51 crc kubenswrapper[4837]: I1001 09:30:51.816974 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:30:51 crc kubenswrapper[4837]: E1001 09:30:51.818113 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:31:02 crc kubenswrapper[4837]: I1001 09:31:02.817799 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:31:03 crc kubenswrapper[4837]: I1001 09:31:03.362622 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217"} Oct 01 09:31:03 crc kubenswrapper[4837]: I1001 09:31:03.365281 4837 generic.go:334] "Generic (PLEG): container finished" podID="b189babd-cae5-47d1-bc90-c8a49033576e" containerID="2015a8484a569162f636303dcaba2f27d418262cb8df9e558c40966750615cbe" exitCode=0 Oct 01 09:31:03 crc kubenswrapper[4837]: I1001 09:31:03.365321 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-547hg" event={"ID":"b189babd-cae5-47d1-bc90-c8a49033576e","Type":"ContainerDied","Data":"2015a8484a569162f636303dcaba2f27d418262cb8df9e558c40966750615cbe"} Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.846069 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.951864 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key\") pod \"b189babd-cae5-47d1-bc90-c8a49033576e\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.951936 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm4d2\" (UniqueName: \"kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2\") pod \"b189babd-cae5-47d1-bc90-c8a49033576e\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.952027 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory\") pod \"b189babd-cae5-47d1-bc90-c8a49033576e\" (UID: \"b189babd-cae5-47d1-bc90-c8a49033576e\") " Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.959187 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2" (OuterVolumeSpecName: "kube-api-access-xm4d2") pod "b189babd-cae5-47d1-bc90-c8a49033576e" (UID: "b189babd-cae5-47d1-bc90-c8a49033576e"). InnerVolumeSpecName "kube-api-access-xm4d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.979883 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b189babd-cae5-47d1-bc90-c8a49033576e" (UID: "b189babd-cae5-47d1-bc90-c8a49033576e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:04 crc kubenswrapper[4837]: I1001 09:31:04.982645 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory" (OuterVolumeSpecName: "inventory") pod "b189babd-cae5-47d1-bc90-c8a49033576e" (UID: "b189babd-cae5-47d1-bc90-c8a49033576e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.054072 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.054105 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm4d2\" (UniqueName: \"kubernetes.io/projected/b189babd-cae5-47d1-bc90-c8a49033576e-kube-api-access-xm4d2\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.054117 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b189babd-cae5-47d1-bc90-c8a49033576e-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.390125 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-547hg" event={"ID":"b189babd-cae5-47d1-bc90-c8a49033576e","Type":"ContainerDied","Data":"4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b"} Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.390558 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cee44de688c351d7799f4827522a07d49ac3fde50f5a7d8c582ad02f03d631b" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.390191 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-547hg" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.538460 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-networker-mfk6z"] Oct 01 09:31:05 crc kubenswrapper[4837]: E1001 09:31:05.539088 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b189babd-cae5-47d1-bc90-c8a49033576e" containerName="configure-network-openstack-openstack-networker" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.539115 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b189babd-cae5-47d1-bc90-c8a49033576e" containerName="configure-network-openstack-openstack-networker" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.539381 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b189babd-cae5-47d1-bc90-c8a49033576e" containerName="configure-network-openstack-openstack-networker" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.540428 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.543603 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.544308 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.560056 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-mfk6z"] Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.678747 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.678916 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d95pk\" (UniqueName: \"kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.678949 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.781863 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.782078 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d95pk\" (UniqueName: \"kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.782122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.791180 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.791434 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.808298 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d95pk\" (UniqueName: \"kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk\") pod \"validate-network-openstack-openstack-networker-mfk6z\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:05 crc kubenswrapper[4837]: I1001 09:31:05.876455 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:06 crc kubenswrapper[4837]: W1001 09:31:06.302195 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4fdd3dd_4022_4ec7_bc01_d489e1127fb0.slice/crio-908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851 WatchSource:0}: Error finding container 908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851: Status 404 returned error can't find the container with id 908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851 Oct 01 09:31:06 crc kubenswrapper[4837]: I1001 09:31:06.304198 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-mfk6z"] Oct 01 09:31:06 crc kubenswrapper[4837]: I1001 09:31:06.309485 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:31:06 crc kubenswrapper[4837]: I1001 09:31:06.402501 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" event={"ID":"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0","Type":"ContainerStarted","Data":"908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851"} Oct 01 09:31:07 crc kubenswrapper[4837]: I1001 09:31:07.416448 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" event={"ID":"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0","Type":"ContainerStarted","Data":"225038973fa9a23ef72b7e5171179c75c7ddc808d113d240e3312e9c60faf251"} Oct 01 09:31:07 crc kubenswrapper[4837]: I1001 09:31:07.442298 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" podStartSLOduration=1.952043274 podStartE2EDuration="2.44227438s" podCreationTimestamp="2025-10-01 09:31:05 +0000 UTC" firstStartedPulling="2025-10-01 09:31:06.309256404 +0000 UTC m=+8723.150863859" lastFinishedPulling="2025-10-01 09:31:06.7994875 +0000 UTC m=+8723.641094965" observedRunningTime="2025-10-01 09:31:07.43374108 +0000 UTC m=+8724.275348545" watchObservedRunningTime="2025-10-01 09:31:07.44227438 +0000 UTC m=+8724.283881875" Oct 01 09:31:13 crc kubenswrapper[4837]: I1001 09:31:13.484592 4837 generic.go:334] "Generic (PLEG): container finished" podID="e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" containerID="225038973fa9a23ef72b7e5171179c75c7ddc808d113d240e3312e9c60faf251" exitCode=0 Oct 01 09:31:13 crc kubenswrapper[4837]: I1001 09:31:13.484680 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" event={"ID":"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0","Type":"ContainerDied","Data":"225038973fa9a23ef72b7e5171179c75c7ddc808d113d240e3312e9c60faf251"} Oct 01 09:31:14 crc kubenswrapper[4837]: I1001 09:31:14.944388 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.107212 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key\") pod \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.107282 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory\") pod \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.107347 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d95pk\" (UniqueName: \"kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk\") pod \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\" (UID: \"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0\") " Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.113636 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk" (OuterVolumeSpecName: "kube-api-access-d95pk") pod "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" (UID: "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0"). InnerVolumeSpecName "kube-api-access-d95pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.145491 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" (UID: "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.157122 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory" (OuterVolumeSpecName: "inventory") pod "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" (UID: "e4fdd3dd-4022-4ec7-bc01-d489e1127fb0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.210687 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.210778 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d95pk\" (UniqueName: \"kubernetes.io/projected/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-kube-api-access-d95pk\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.210800 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4fdd3dd-4022-4ec7-bc01-d489e1127fb0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.515099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" event={"ID":"e4fdd3dd-4022-4ec7-bc01-d489e1127fb0","Type":"ContainerDied","Data":"908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851"} Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.515171 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="908c9fc4cc2827fc55786065089d12a7c7c5b18cb72a70c46dffe48541d5e851" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.515184 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-mfk6z" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.572880 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-networker-gsgxm"] Oct 01 09:31:15 crc kubenswrapper[4837]: E1001 09:31:15.576005 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" containerName="validate-network-openstack-openstack-networker" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.576049 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" containerName="validate-network-openstack-openstack-networker" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.576322 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4fdd3dd-4022-4ec7-bc01-d489e1127fb0" containerName="validate-network-openstack-openstack-networker" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.578792 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.587564 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.588613 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.594239 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-gsgxm"] Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.720327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.720840 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7k9q\" (UniqueName: \"kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.721141 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.822543 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.822616 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.822684 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7k9q\" (UniqueName: \"kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.829452 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.830550 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.841931 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7k9q\" (UniqueName: \"kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q\") pod \"install-os-openstack-openstack-networker-gsgxm\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:15 crc kubenswrapper[4837]: I1001 09:31:15.914578 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:31:16 crc kubenswrapper[4837]: I1001 09:31:16.309063 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-gsgxm"] Oct 01 09:31:16 crc kubenswrapper[4837]: W1001 09:31:16.314503 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89f2b30f_c5e3_4a62_9092_4b00ea001f3c.slice/crio-59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678 WatchSource:0}: Error finding container 59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678: Status 404 returned error can't find the container with id 59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678 Oct 01 09:31:16 crc kubenswrapper[4837]: I1001 09:31:16.524707 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-gsgxm" event={"ID":"89f2b30f-c5e3-4a62-9092-4b00ea001f3c","Type":"ContainerStarted","Data":"59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678"} Oct 01 09:31:17 crc kubenswrapper[4837]: I1001 09:31:17.536422 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-gsgxm" event={"ID":"89f2b30f-c5e3-4a62-9092-4b00ea001f3c","Type":"ContainerStarted","Data":"4b06757b6f5d42d888bc38dec70180ed9de1e8d856afa6e06a65dbcfd862fa3b"} Oct 01 09:31:17 crc kubenswrapper[4837]: I1001 09:31:17.569056 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-networker-gsgxm" podStartSLOduration=2.043798896 podStartE2EDuration="2.569036312s" podCreationTimestamp="2025-10-01 09:31:15 +0000 UTC" firstStartedPulling="2025-10-01 09:31:16.318315441 +0000 UTC m=+8733.159922896" lastFinishedPulling="2025-10-01 09:31:16.843552817 +0000 UTC m=+8733.685160312" observedRunningTime="2025-10-01 09:31:17.562554483 +0000 UTC m=+8734.404161938" watchObservedRunningTime="2025-10-01 09:31:17.569036312 +0000 UTC m=+8734.410643777" Oct 01 09:31:24 crc kubenswrapper[4837]: I1001 09:31:24.630786 4837 generic.go:334] "Generic (PLEG): container finished" podID="9ccf0a84-b87b-4a54-b268-8b55d19dba49" containerID="b02a0dace5b2b9051452b9bbb0dcf71f05042e9ac99c54c79c701b53333d2784" exitCode=0 Oct 01 09:31:24 crc kubenswrapper[4837]: I1001 09:31:24.631143 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" event={"ID":"9ccf0a84-b87b-4a54-b268-8b55d19dba49","Type":"ContainerDied","Data":"b02a0dace5b2b9051452b9bbb0dcf71f05042e9ac99c54c79c701b53333d2784"} Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.118671 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.252774 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hvtv\" (UniqueName: \"kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv\") pod \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.253115 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory\") pod \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.253227 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") pod \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.263160 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv" (OuterVolumeSpecName: "kube-api-access-6hvtv") pod "9ccf0a84-b87b-4a54-b268-8b55d19dba49" (UID: "9ccf0a84-b87b-4a54-b268-8b55d19dba49"). InnerVolumeSpecName "kube-api-access-6hvtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:31:26 crc kubenswrapper[4837]: E1001 09:31:26.304014 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key podName:9ccf0a84-b87b-4a54-b268-8b55d19dba49 nodeName:}" failed. No retries permitted until 2025-10-01 09:31:26.803976972 +0000 UTC m=+8743.645584477 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key") pod "9ccf0a84-b87b-4a54-b268-8b55d19dba49" (UID: "9ccf0a84-b87b-4a54-b268-8b55d19dba49") : error deleting /var/lib/kubelet/pods/9ccf0a84-b87b-4a54-b268-8b55d19dba49/volume-subpaths: remove /var/lib/kubelet/pods/9ccf0a84-b87b-4a54-b268-8b55d19dba49/volume-subpaths: no such file or directory Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.307975 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory" (OuterVolumeSpecName: "inventory") pod "9ccf0a84-b87b-4a54-b268-8b55d19dba49" (UID: "9ccf0a84-b87b-4a54-b268-8b55d19dba49"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.355975 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hvtv\" (UniqueName: \"kubernetes.io/projected/9ccf0a84-b87b-4a54-b268-8b55d19dba49-kube-api-access-6hvtv\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.356152 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.654807 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" event={"ID":"9ccf0a84-b87b-4a54-b268-8b55d19dba49","Type":"ContainerDied","Data":"63257ddc5118bbf9f674d59c9cbeeb699846b87a6b57e31855320ceafa9066b5"} Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.654860 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63257ddc5118bbf9f674d59c9cbeeb699846b87a6b57e31855320ceafa9066b5" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.655066 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-hpq4d" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.762047 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-7nqgj"] Oct 01 09:31:26 crc kubenswrapper[4837]: E1001 09:31:26.764482 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf0a84-b87b-4a54-b268-8b55d19dba49" containerName="configure-network-openstack-openstack-cell1" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.764507 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf0a84-b87b-4a54-b268-8b55d19dba49" containerName="configure-network-openstack-openstack-cell1" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.764772 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ccf0a84-b87b-4a54-b268-8b55d19dba49" containerName="configure-network-openstack-openstack-cell1" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.766591 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.781521 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-7nqgj"] Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.866545 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") pod \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\" (UID: \"9ccf0a84-b87b-4a54-b268-8b55d19dba49\") " Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.867041 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.867131 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.867424 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8956\" (UniqueName: \"kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.869666 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ccf0a84-b87b-4a54-b268-8b55d19dba49" (UID: "9ccf0a84-b87b-4a54-b268-8b55d19dba49"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.970088 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8956\" (UniqueName: \"kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.970229 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.970258 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.970337 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ccf0a84-b87b-4a54-b268-8b55d19dba49-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.973148 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.973264 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:26 crc kubenswrapper[4837]: I1001 09:31:26.987607 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8956\" (UniqueName: \"kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956\") pod \"validate-network-openstack-openstack-cell1-7nqgj\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:27 crc kubenswrapper[4837]: I1001 09:31:27.095832 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:27 crc kubenswrapper[4837]: I1001 09:31:27.704158 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-7nqgj"] Oct 01 09:31:28 crc kubenswrapper[4837]: I1001 09:31:28.669736 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" event={"ID":"85317dc9-c42f-421a-9101-787a19352703","Type":"ContainerStarted","Data":"1ec3631ca9ae04218f289ab4b596e23a0c51cfadb06e5d6edb94cdf1d9491503"} Oct 01 09:31:28 crc kubenswrapper[4837]: I1001 09:31:28.670045 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" event={"ID":"85317dc9-c42f-421a-9101-787a19352703","Type":"ContainerStarted","Data":"944c1df5d0ab2598abea3935d29209bd71f09b41b553f32edb98dac3693fa02f"} Oct 01 09:31:28 crc kubenswrapper[4837]: I1001 09:31:28.695627 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" podStartSLOduration=2.241498037 podStartE2EDuration="2.695604593s" podCreationTimestamp="2025-10-01 09:31:26 +0000 UTC" firstStartedPulling="2025-10-01 09:31:27.706405238 +0000 UTC m=+8744.548012693" lastFinishedPulling="2025-10-01 09:31:28.160511784 +0000 UTC m=+8745.002119249" observedRunningTime="2025-10-01 09:31:28.687144765 +0000 UTC m=+8745.528752220" watchObservedRunningTime="2025-10-01 09:31:28.695604593 +0000 UTC m=+8745.537212048" Oct 01 09:31:33 crc kubenswrapper[4837]: I1001 09:31:33.717365 4837 generic.go:334] "Generic (PLEG): container finished" podID="85317dc9-c42f-421a-9101-787a19352703" containerID="1ec3631ca9ae04218f289ab4b596e23a0c51cfadb06e5d6edb94cdf1d9491503" exitCode=0 Oct 01 09:31:33 crc kubenswrapper[4837]: I1001 09:31:33.717495 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" event={"ID":"85317dc9-c42f-421a-9101-787a19352703","Type":"ContainerDied","Data":"1ec3631ca9ae04218f289ab4b596e23a0c51cfadb06e5d6edb94cdf1d9491503"} Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.333769 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.450386 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory\") pod \"85317dc9-c42f-421a-9101-787a19352703\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.450511 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key\") pod \"85317dc9-c42f-421a-9101-787a19352703\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.450852 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8956\" (UniqueName: \"kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956\") pod \"85317dc9-c42f-421a-9101-787a19352703\" (UID: \"85317dc9-c42f-421a-9101-787a19352703\") " Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.456309 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956" (OuterVolumeSpecName: "kube-api-access-r8956") pod "85317dc9-c42f-421a-9101-787a19352703" (UID: "85317dc9-c42f-421a-9101-787a19352703"). InnerVolumeSpecName "kube-api-access-r8956". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.487110 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85317dc9-c42f-421a-9101-787a19352703" (UID: "85317dc9-c42f-421a-9101-787a19352703"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.493042 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory" (OuterVolumeSpecName: "inventory") pod "85317dc9-c42f-421a-9101-787a19352703" (UID: "85317dc9-c42f-421a-9101-787a19352703"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.553554 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8956\" (UniqueName: \"kubernetes.io/projected/85317dc9-c42f-421a-9101-787a19352703-kube-api-access-r8956\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.553589 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.553600 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85317dc9-c42f-421a-9101-787a19352703-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.742540 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" event={"ID":"85317dc9-c42f-421a-9101-787a19352703","Type":"ContainerDied","Data":"944c1df5d0ab2598abea3935d29209bd71f09b41b553f32edb98dac3693fa02f"} Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.742580 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="944c1df5d0ab2598abea3935d29209bd71f09b41b553f32edb98dac3693fa02f" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.742618 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-7nqgj" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.848068 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-5bp2d"] Oct 01 09:31:35 crc kubenswrapper[4837]: E1001 09:31:35.848667 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85317dc9-c42f-421a-9101-787a19352703" containerName="validate-network-openstack-openstack-cell1" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.848704 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="85317dc9-c42f-421a-9101-787a19352703" containerName="validate-network-openstack-openstack-cell1" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.848928 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="85317dc9-c42f-421a-9101-787a19352703" containerName="validate-network-openstack-openstack-cell1" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.849868 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.852538 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.855445 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.856776 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-5bp2d"] Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.961852 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.961936 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgt4r\" (UniqueName: \"kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:35 crc kubenswrapper[4837]: I1001 09:31:35.962019 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.063575 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.063861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.064656 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgt4r\" (UniqueName: \"kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.068222 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.069183 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.098880 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgt4r\" (UniqueName: \"kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r\") pod \"install-os-openstack-openstack-cell1-5bp2d\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.185960 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:31:36 crc kubenswrapper[4837]: I1001 09:31:36.754225 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-5bp2d"] Oct 01 09:31:37 crc kubenswrapper[4837]: I1001 09:31:37.768320 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" event={"ID":"5040522c-817c-455f-a58c-5becbf01ac90","Type":"ContainerStarted","Data":"8c60087f4202d5f3a08edf1668b96e13a21a41d3f6a3a7b8ba4fc567f8e51c4a"} Oct 01 09:31:37 crc kubenswrapper[4837]: I1001 09:31:37.768641 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" event={"ID":"5040522c-817c-455f-a58c-5becbf01ac90","Type":"ContainerStarted","Data":"97d651756bc29745a7038c79f68c807dbbcd16a4b965482d8f186e6a9832dc6d"} Oct 01 09:31:37 crc kubenswrapper[4837]: I1001 09:31:37.783392 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" podStartSLOduration=2.285835671 podStartE2EDuration="2.783365896s" podCreationTimestamp="2025-10-01 09:31:35 +0000 UTC" firstStartedPulling="2025-10-01 09:31:36.765550406 +0000 UTC m=+8753.607157871" lastFinishedPulling="2025-10-01 09:31:37.263080621 +0000 UTC m=+8754.104688096" observedRunningTime="2025-10-01 09:31:37.782021783 +0000 UTC m=+8754.623629268" watchObservedRunningTime="2025-10-01 09:31:37.783365896 +0000 UTC m=+8754.624973391" Oct 01 09:32:04 crc kubenswrapper[4837]: I1001 09:32:04.074435 4837 generic.go:334] "Generic (PLEG): container finished" podID="89f2b30f-c5e3-4a62-9092-4b00ea001f3c" containerID="4b06757b6f5d42d888bc38dec70180ed9de1e8d856afa6e06a65dbcfd862fa3b" exitCode=0 Oct 01 09:32:04 crc kubenswrapper[4837]: I1001 09:32:04.074566 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-gsgxm" event={"ID":"89f2b30f-c5e3-4a62-9092-4b00ea001f3c","Type":"ContainerDied","Data":"4b06757b6f5d42d888bc38dec70180ed9de1e8d856afa6e06a65dbcfd862fa3b"} Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.633389 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.700718 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory\") pod \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.700804 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key\") pod \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.700839 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7k9q\" (UniqueName: \"kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q\") pod \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\" (UID: \"89f2b30f-c5e3-4a62-9092-4b00ea001f3c\") " Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.712849 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q" (OuterVolumeSpecName: "kube-api-access-v7k9q") pod "89f2b30f-c5e3-4a62-9092-4b00ea001f3c" (UID: "89f2b30f-c5e3-4a62-9092-4b00ea001f3c"). InnerVolumeSpecName "kube-api-access-v7k9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.735251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "89f2b30f-c5e3-4a62-9092-4b00ea001f3c" (UID: "89f2b30f-c5e3-4a62-9092-4b00ea001f3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.744191 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory" (OuterVolumeSpecName: "inventory") pod "89f2b30f-c5e3-4a62-9092-4b00ea001f3c" (UID: "89f2b30f-c5e3-4a62-9092-4b00ea001f3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.803583 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.803617 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:05 crc kubenswrapper[4837]: I1001 09:32:05.803629 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7k9q\" (UniqueName: \"kubernetes.io/projected/89f2b30f-c5e3-4a62-9092-4b00ea001f3c-kube-api-access-v7k9q\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.104169 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-gsgxm" event={"ID":"89f2b30f-c5e3-4a62-9092-4b00ea001f3c","Type":"ContainerDied","Data":"59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678"} Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.104231 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59cd6849b9b437366caf20b27da1bd5d60a50e917e144d326f9e3f9e963bb678" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.104244 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-gsgxm" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.219669 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4qf6g"] Oct 01 09:32:06 crc kubenswrapper[4837]: E1001 09:32:06.220299 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f2b30f-c5e3-4a62-9092-4b00ea001f3c" containerName="install-os-openstack-openstack-networker" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.220322 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f2b30f-c5e3-4a62-9092-4b00ea001f3c" containerName="install-os-openstack-openstack-networker" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.220590 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="89f2b30f-c5e3-4a62-9092-4b00ea001f3c" containerName="install-os-openstack-openstack-networker" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.221649 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.224585 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.224987 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.238768 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4qf6g"] Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.313650 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j5f4\" (UniqueName: \"kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.313865 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.314049 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.416649 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.416798 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.416890 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j5f4\" (UniqueName: \"kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.422056 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.422283 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.459372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j5f4\" (UniqueName: \"kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4\") pod \"configure-os-openstack-openstack-networker-4qf6g\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:06 crc kubenswrapper[4837]: I1001 09:32:06.540544 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:32:07 crc kubenswrapper[4837]: I1001 09:32:07.152289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-4qf6g"] Oct 01 09:32:08 crc kubenswrapper[4837]: I1001 09:32:08.121772 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" event={"ID":"36a513c6-8a4c-4f84-b622-8e29610038ff","Type":"ContainerStarted","Data":"1762bbec672e52b36c3b79ed029c3d8eebbc85a9d498a78ced7b3874a2e8b439"} Oct 01 09:32:08 crc kubenswrapper[4837]: I1001 09:32:08.122054 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" event={"ID":"36a513c6-8a4c-4f84-b622-8e29610038ff","Type":"ContainerStarted","Data":"057cff0aa90864c735137a8d017acaef3169a51bab4715dc299227c30fb7f4ef"} Oct 01 09:32:08 crc kubenswrapper[4837]: I1001 09:32:08.158587 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" podStartSLOduration=1.525622204 podStartE2EDuration="2.158561311s" podCreationTimestamp="2025-10-01 09:32:06 +0000 UTC" firstStartedPulling="2025-10-01 09:32:07.144962795 +0000 UTC m=+8783.986570260" lastFinishedPulling="2025-10-01 09:32:07.777901872 +0000 UTC m=+8784.619509367" observedRunningTime="2025-10-01 09:32:08.153938368 +0000 UTC m=+8784.995545833" watchObservedRunningTime="2025-10-01 09:32:08.158561311 +0000 UTC m=+8785.000168766" Oct 01 09:32:26 crc kubenswrapper[4837]: I1001 09:32:26.329565 4837 generic.go:334] "Generic (PLEG): container finished" podID="5040522c-817c-455f-a58c-5becbf01ac90" containerID="8c60087f4202d5f3a08edf1668b96e13a21a41d3f6a3a7b8ba4fc567f8e51c4a" exitCode=0 Oct 01 09:32:26 crc kubenswrapper[4837]: I1001 09:32:26.329682 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" event={"ID":"5040522c-817c-455f-a58c-5becbf01ac90","Type":"ContainerDied","Data":"8c60087f4202d5f3a08edf1668b96e13a21a41d3f6a3a7b8ba4fc567f8e51c4a"} Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.361088 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" event={"ID":"5040522c-817c-455f-a58c-5becbf01ac90","Type":"ContainerDied","Data":"97d651756bc29745a7038c79f68c807dbbcd16a4b965482d8f186e6a9832dc6d"} Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.361902 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97d651756bc29745a7038c79f68c807dbbcd16a4b965482d8f186e6a9832dc6d" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.414972 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.548648 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory\") pod \"5040522c-817c-455f-a58c-5becbf01ac90\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.548753 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key\") pod \"5040522c-817c-455f-a58c-5becbf01ac90\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.548907 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgt4r\" (UniqueName: \"kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r\") pod \"5040522c-817c-455f-a58c-5becbf01ac90\" (UID: \"5040522c-817c-455f-a58c-5becbf01ac90\") " Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.556546 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r" (OuterVolumeSpecName: "kube-api-access-jgt4r") pod "5040522c-817c-455f-a58c-5becbf01ac90" (UID: "5040522c-817c-455f-a58c-5becbf01ac90"). InnerVolumeSpecName "kube-api-access-jgt4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.608586 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory" (OuterVolumeSpecName: "inventory") pod "5040522c-817c-455f-a58c-5becbf01ac90" (UID: "5040522c-817c-455f-a58c-5becbf01ac90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.625886 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5040522c-817c-455f-a58c-5becbf01ac90" (UID: "5040522c-817c-455f-a58c-5becbf01ac90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.652331 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgt4r\" (UniqueName: \"kubernetes.io/projected/5040522c-817c-455f-a58c-5becbf01ac90-kube-api-access-jgt4r\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.652365 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:28 crc kubenswrapper[4837]: I1001 09:32:28.652377 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5040522c-817c-455f-a58c-5becbf01ac90-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.373821 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-5bp2d" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.540414 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hsszd"] Oct 01 09:32:29 crc kubenswrapper[4837]: E1001 09:32:29.540813 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5040522c-817c-455f-a58c-5becbf01ac90" containerName="install-os-openstack-openstack-cell1" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.540826 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5040522c-817c-455f-a58c-5becbf01ac90" containerName="install-os-openstack-openstack-cell1" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.541029 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5040522c-817c-455f-a58c-5becbf01ac90" containerName="install-os-openstack-openstack-cell1" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.541725 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.543818 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.545318 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.555375 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hsszd"] Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.674962 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgcf\" (UniqueName: \"kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.675018 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.675131 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.799931 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgcf\" (UniqueName: \"kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.799993 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.800095 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.808576 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.812763 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.823935 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgcf\" (UniqueName: \"kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf\") pod \"configure-os-openstack-openstack-cell1-hsszd\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:29 crc kubenswrapper[4837]: I1001 09:32:29.860058 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:32:30 crc kubenswrapper[4837]: I1001 09:32:30.468169 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hsszd"] Oct 01 09:32:30 crc kubenswrapper[4837]: W1001 09:32:30.475891 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5fea961_8894_40a6_a2e6_31d5b2646323.slice/crio-ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e WatchSource:0}: Error finding container ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e: Status 404 returned error can't find the container with id ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e Oct 01 09:32:31 crc kubenswrapper[4837]: I1001 09:32:31.399942 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" event={"ID":"d5fea961-8894-40a6-a2e6-31d5b2646323","Type":"ContainerStarted","Data":"d8c2a5e8a00613fe8e1602696ffd272ef5cea428b17b4bf5583df2b66b021df4"} Oct 01 09:32:31 crc kubenswrapper[4837]: I1001 09:32:31.400366 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" event={"ID":"d5fea961-8894-40a6-a2e6-31d5b2646323","Type":"ContainerStarted","Data":"ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e"} Oct 01 09:32:31 crc kubenswrapper[4837]: I1001 09:32:31.435608 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" podStartSLOduration=2.018327862 podStartE2EDuration="2.435583271s" podCreationTimestamp="2025-10-01 09:32:29 +0000 UTC" firstStartedPulling="2025-10-01 09:32:30.478203099 +0000 UTC m=+8807.319810564" lastFinishedPulling="2025-10-01 09:32:30.895458518 +0000 UTC m=+8807.737065973" observedRunningTime="2025-10-01 09:32:31.434524645 +0000 UTC m=+8808.276132110" watchObservedRunningTime="2025-10-01 09:32:31.435583271 +0000 UTC m=+8808.277190756" Oct 01 09:33:02 crc kubenswrapper[4837]: I1001 09:33:02.819941 4837 generic.go:334] "Generic (PLEG): container finished" podID="36a513c6-8a4c-4f84-b622-8e29610038ff" containerID="1762bbec672e52b36c3b79ed029c3d8eebbc85a9d498a78ced7b3874a2e8b439" exitCode=2 Oct 01 09:33:02 crc kubenswrapper[4837]: I1001 09:33:02.820132 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" event={"ID":"36a513c6-8a4c-4f84-b622-8e29610038ff","Type":"ContainerDied","Data":"1762bbec672e52b36c3b79ed029c3d8eebbc85a9d498a78ced7b3874a2e8b439"} Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.354800 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.403807 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j5f4\" (UniqueName: \"kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4\") pod \"36a513c6-8a4c-4f84-b622-8e29610038ff\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.404101 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory\") pod \"36a513c6-8a4c-4f84-b622-8e29610038ff\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.404219 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key\") pod \"36a513c6-8a4c-4f84-b622-8e29610038ff\" (UID: \"36a513c6-8a4c-4f84-b622-8e29610038ff\") " Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.412179 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4" (OuterVolumeSpecName: "kube-api-access-8j5f4") pod "36a513c6-8a4c-4f84-b622-8e29610038ff" (UID: "36a513c6-8a4c-4f84-b622-8e29610038ff"). InnerVolumeSpecName "kube-api-access-8j5f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.436241 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory" (OuterVolumeSpecName: "inventory") pod "36a513c6-8a4c-4f84-b622-8e29610038ff" (UID: "36a513c6-8a4c-4f84-b622-8e29610038ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.455101 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "36a513c6-8a4c-4f84-b622-8e29610038ff" (UID: "36a513c6-8a4c-4f84-b622-8e29610038ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.507085 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.507116 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36a513c6-8a4c-4f84-b622-8e29610038ff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.507126 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j5f4\" (UniqueName: \"kubernetes.io/projected/36a513c6-8a4c-4f84-b622-8e29610038ff-kube-api-access-8j5f4\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.849267 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" event={"ID":"36a513c6-8a4c-4f84-b622-8e29610038ff","Type":"ContainerDied","Data":"057cff0aa90864c735137a8d017acaef3169a51bab4715dc299227c30fb7f4ef"} Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.849331 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="057cff0aa90864c735137a8d017acaef3169a51bab4715dc299227c30fb7f4ef" Oct 01 09:33:04 crc kubenswrapper[4837]: I1001 09:33:04.849358 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-4qf6g" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.044012 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-snpwc"] Oct 01 09:33:12 crc kubenswrapper[4837]: E1001 09:33:12.045666 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a513c6-8a4c-4f84-b622-8e29610038ff" containerName="configure-os-openstack-openstack-networker" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.045742 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a513c6-8a4c-4f84-b622-8e29610038ff" containerName="configure-os-openstack-openstack-networker" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.046304 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a513c6-8a4c-4f84-b622-8e29610038ff" containerName="configure-os-openstack-openstack-networker" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.048086 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.053206 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.053795 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.064207 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-snpwc"] Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.096957 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.097411 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg7kb\" (UniqueName: \"kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.097538 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.198626 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg7kb\" (UniqueName: \"kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.198739 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.198813 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.207298 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.207382 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.214327 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg7kb\" (UniqueName: \"kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb\") pod \"configure-os-openstack-openstack-networker-snpwc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:12 crc kubenswrapper[4837]: I1001 09:33:12.413064 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:33:13 crc kubenswrapper[4837]: I1001 09:33:13.052181 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-snpwc"] Oct 01 09:33:13 crc kubenswrapper[4837]: W1001 09:33:13.056976 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9e207f3_0eca_47c1_aebf_de892fe662cc.slice/crio-dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22 WatchSource:0}: Error finding container dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22: Status 404 returned error can't find the container with id dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22 Oct 01 09:33:13 crc kubenswrapper[4837]: I1001 09:33:13.971494 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-snpwc" event={"ID":"d9e207f3-0eca-47c1-aebf-de892fe662cc","Type":"ContainerStarted","Data":"452f56ab679013eb48ca69c0efdf0d2379928d43c4c7dd6c80cf94432549f914"} Oct 01 09:33:13 crc kubenswrapper[4837]: I1001 09:33:13.971888 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-snpwc" event={"ID":"d9e207f3-0eca-47c1-aebf-de892fe662cc","Type":"ContainerStarted","Data":"dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22"} Oct 01 09:33:14 crc kubenswrapper[4837]: I1001 09:33:14.024387 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-snpwc" podStartSLOduration=1.473036649 podStartE2EDuration="2.024355198s" podCreationTimestamp="2025-10-01 09:33:12 +0000 UTC" firstStartedPulling="2025-10-01 09:33:13.0594338 +0000 UTC m=+8849.901041265" lastFinishedPulling="2025-10-01 09:33:13.610752319 +0000 UTC m=+8850.452359814" observedRunningTime="2025-10-01 09:33:13.984443696 +0000 UTC m=+8850.826051161" watchObservedRunningTime="2025-10-01 09:33:14.024355198 +0000 UTC m=+8850.865962683" Oct 01 09:33:23 crc kubenswrapper[4837]: I1001 09:33:23.080196 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:33:23 crc kubenswrapper[4837]: I1001 09:33:23.080976 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:33:25 crc kubenswrapper[4837]: I1001 09:33:25.113250 4837 generic.go:334] "Generic (PLEG): container finished" podID="d5fea961-8894-40a6-a2e6-31d5b2646323" containerID="d8c2a5e8a00613fe8e1602696ffd272ef5cea428b17b4bf5583df2b66b021df4" exitCode=2 Oct 01 09:33:25 crc kubenswrapper[4837]: I1001 09:33:25.113386 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" event={"ID":"d5fea961-8894-40a6-a2e6-31d5b2646323","Type":"ContainerDied","Data":"d8c2a5e8a00613fe8e1602696ffd272ef5cea428b17b4bf5583df2b66b021df4"} Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.615444 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.734605 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key\") pod \"d5fea961-8894-40a6-a2e6-31d5b2646323\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.734663 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltgcf\" (UniqueName: \"kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf\") pod \"d5fea961-8894-40a6-a2e6-31d5b2646323\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.734885 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory\") pod \"d5fea961-8894-40a6-a2e6-31d5b2646323\" (UID: \"d5fea961-8894-40a6-a2e6-31d5b2646323\") " Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.742665 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf" (OuterVolumeSpecName: "kube-api-access-ltgcf") pod "d5fea961-8894-40a6-a2e6-31d5b2646323" (UID: "d5fea961-8894-40a6-a2e6-31d5b2646323"). InnerVolumeSpecName "kube-api-access-ltgcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.779038 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5fea961-8894-40a6-a2e6-31d5b2646323" (UID: "d5fea961-8894-40a6-a2e6-31d5b2646323"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.783950 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory" (OuterVolumeSpecName: "inventory") pod "d5fea961-8894-40a6-a2e6-31d5b2646323" (UID: "d5fea961-8894-40a6-a2e6-31d5b2646323"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.837627 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.837809 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltgcf\" (UniqueName: \"kubernetes.io/projected/d5fea961-8894-40a6-a2e6-31d5b2646323-kube-api-access-ltgcf\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:26 crc kubenswrapper[4837]: I1001 09:33:26.837888 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fea961-8894-40a6-a2e6-31d5b2646323-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:33:27 crc kubenswrapper[4837]: I1001 09:33:27.140469 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" event={"ID":"d5fea961-8894-40a6-a2e6-31d5b2646323","Type":"ContainerDied","Data":"ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e"} Oct 01 09:33:27 crc kubenswrapper[4837]: I1001 09:33:27.140914 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba852b4ffdf8f34a11fad23d2af7b1882a2060d556abcedc9c92ecb65f11173e" Oct 01 09:33:27 crc kubenswrapper[4837]: I1001 09:33:27.140544 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hsszd" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.074034 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gmdpl"] Oct 01 09:33:34 crc kubenswrapper[4837]: E1001 09:33:34.079482 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5fea961-8894-40a6-a2e6-31d5b2646323" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.079618 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5fea961-8894-40a6-a2e6-31d5b2646323" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.080713 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5fea961-8894-40a6-a2e6-31d5b2646323" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.082678 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.096980 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gmdpl"] Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.099388 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.099571 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.213910 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df7sv\" (UniqueName: \"kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.213969 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.214097 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.315439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.315603 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.315751 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df7sv\" (UniqueName: \"kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.327433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.327803 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.346584 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df7sv\" (UniqueName: \"kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv\") pod \"configure-os-openstack-openstack-cell1-gmdpl\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:34 crc kubenswrapper[4837]: I1001 09:33:34.423809 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:33:35 crc kubenswrapper[4837]: I1001 09:33:35.024207 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gmdpl"] Oct 01 09:33:35 crc kubenswrapper[4837]: I1001 09:33:35.234253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" event={"ID":"587efc59-4844-4bb9-8467-5c39b985fab0","Type":"ContainerStarted","Data":"f3ba83d011e733d212a3aedbf62a790fd0095c9ea44168e898810746c58fdc10"} Oct 01 09:33:36 crc kubenswrapper[4837]: I1001 09:33:36.250742 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" event={"ID":"587efc59-4844-4bb9-8467-5c39b985fab0","Type":"ContainerStarted","Data":"5f3094440856cf18259d9ed26882868f15fe55a0400c48d43ca590a193a8ed1f"} Oct 01 09:33:36 crc kubenswrapper[4837]: I1001 09:33:36.287134 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" podStartSLOduration=1.849176818 podStartE2EDuration="2.287116186s" podCreationTimestamp="2025-10-01 09:33:34 +0000 UTC" firstStartedPulling="2025-10-01 09:33:35.015169301 +0000 UTC m=+8871.856776766" lastFinishedPulling="2025-10-01 09:33:35.453108679 +0000 UTC m=+8872.294716134" observedRunningTime="2025-10-01 09:33:36.284117992 +0000 UTC m=+8873.125725447" watchObservedRunningTime="2025-10-01 09:33:36.287116186 +0000 UTC m=+8873.128723641" Oct 01 09:33:53 crc kubenswrapper[4837]: I1001 09:33:53.079016 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:33:53 crc kubenswrapper[4837]: I1001 09:33:53.079565 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:34:02 crc kubenswrapper[4837]: I1001 09:34:02.569060 4837 generic.go:334] "Generic (PLEG): container finished" podID="d9e207f3-0eca-47c1-aebf-de892fe662cc" containerID="452f56ab679013eb48ca69c0efdf0d2379928d43c4c7dd6c80cf94432549f914" exitCode=0 Oct 01 09:34:02 crc kubenswrapper[4837]: I1001 09:34:02.569171 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-snpwc" event={"ID":"d9e207f3-0eca-47c1-aebf-de892fe662cc","Type":"ContainerDied","Data":"452f56ab679013eb48ca69c0efdf0d2379928d43c4c7dd6c80cf94432549f914"} Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.055476 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.236383 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key\") pod \"d9e207f3-0eca-47c1-aebf-de892fe662cc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.236833 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory\") pod \"d9e207f3-0eca-47c1-aebf-de892fe662cc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.237076 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg7kb\" (UniqueName: \"kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb\") pod \"d9e207f3-0eca-47c1-aebf-de892fe662cc\" (UID: \"d9e207f3-0eca-47c1-aebf-de892fe662cc\") " Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.243177 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb" (OuterVolumeSpecName: "kube-api-access-tg7kb") pod "d9e207f3-0eca-47c1-aebf-de892fe662cc" (UID: "d9e207f3-0eca-47c1-aebf-de892fe662cc"). InnerVolumeSpecName "kube-api-access-tg7kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.272649 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d9e207f3-0eca-47c1-aebf-de892fe662cc" (UID: "d9e207f3-0eca-47c1-aebf-de892fe662cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.273977 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory" (OuterVolumeSpecName: "inventory") pod "d9e207f3-0eca-47c1-aebf-de892fe662cc" (UID: "d9e207f3-0eca-47c1-aebf-de892fe662cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.339893 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.339932 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9e207f3-0eca-47c1-aebf-de892fe662cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.339941 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg7kb\" (UniqueName: \"kubernetes.io/projected/d9e207f3-0eca-47c1-aebf-de892fe662cc-kube-api-access-tg7kb\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.594786 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-snpwc" event={"ID":"d9e207f3-0eca-47c1-aebf-de892fe662cc","Type":"ContainerDied","Data":"dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22"} Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.595160 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc2b261e80052e1267fe21a08b9e6c081d25d4bcf539a539f4c2e94f3a944f22" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.594881 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-snpwc" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.722282 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-networker-jffww"] Oct 01 09:34:04 crc kubenswrapper[4837]: E1001 09:34:04.723296 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e207f3-0eca-47c1-aebf-de892fe662cc" containerName="configure-os-openstack-openstack-networker" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.723456 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e207f3-0eca-47c1-aebf-de892fe662cc" containerName="configure-os-openstack-openstack-networker" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.724126 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e207f3-0eca-47c1-aebf-de892fe662cc" containerName="configure-os-openstack-openstack-networker" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.725607 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.729716 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.729956 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.739475 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-jffww"] Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.861542 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldqgc\" (UniqueName: \"kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.861719 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.861845 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.963952 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldqgc\" (UniqueName: \"kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.964064 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.964173 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.968845 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.970781 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:04 crc kubenswrapper[4837]: I1001 09:34:04.995439 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldqgc\" (UniqueName: \"kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc\") pod \"run-os-openstack-openstack-networker-jffww\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:05 crc kubenswrapper[4837]: I1001 09:34:05.078960 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:05 crc kubenswrapper[4837]: I1001 09:34:05.643809 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-jffww"] Oct 01 09:34:06 crc kubenswrapper[4837]: I1001 09:34:06.628898 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-jffww" event={"ID":"5f6065c8-1d4b-4302-84e4-7e30860f1864","Type":"ContainerStarted","Data":"45cc108eb187ffc9815f21d00ad659a21939e291fba7e9f209c49dc527cf898a"} Oct 01 09:34:06 crc kubenswrapper[4837]: I1001 09:34:06.629341 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-jffww" event={"ID":"5f6065c8-1d4b-4302-84e4-7e30860f1864","Type":"ContainerStarted","Data":"9bd0b1f07c094acd9a36ca3424c1716e98d66107fcd148b01029a92a39bbd2a4"} Oct 01 09:34:06 crc kubenswrapper[4837]: I1001 09:34:06.677545 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-networker-jffww" podStartSLOduration=2.17754794 podStartE2EDuration="2.677522455s" podCreationTimestamp="2025-10-01 09:34:04 +0000 UTC" firstStartedPulling="2025-10-01 09:34:05.665301173 +0000 UTC m=+8902.506908628" lastFinishedPulling="2025-10-01 09:34:06.165275648 +0000 UTC m=+8903.006883143" observedRunningTime="2025-10-01 09:34:06.654851387 +0000 UTC m=+8903.496458882" watchObservedRunningTime="2025-10-01 09:34:06.677522455 +0000 UTC m=+8903.519129920" Oct 01 09:34:14 crc kubenswrapper[4837]: I1001 09:34:14.711782 4837 generic.go:334] "Generic (PLEG): container finished" podID="5f6065c8-1d4b-4302-84e4-7e30860f1864" containerID="45cc108eb187ffc9815f21d00ad659a21939e291fba7e9f209c49dc527cf898a" exitCode=0 Oct 01 09:34:14 crc kubenswrapper[4837]: I1001 09:34:14.711883 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-jffww" event={"ID":"5f6065c8-1d4b-4302-84e4-7e30860f1864","Type":"ContainerDied","Data":"45cc108eb187ffc9815f21d00ad659a21939e291fba7e9f209c49dc527cf898a"} Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.248548 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.349604 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key\") pod \"5f6065c8-1d4b-4302-84e4-7e30860f1864\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.349803 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory\") pod \"5f6065c8-1d4b-4302-84e4-7e30860f1864\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.349878 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldqgc\" (UniqueName: \"kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc\") pod \"5f6065c8-1d4b-4302-84e4-7e30860f1864\" (UID: \"5f6065c8-1d4b-4302-84e4-7e30860f1864\") " Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.356333 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc" (OuterVolumeSpecName: "kube-api-access-ldqgc") pod "5f6065c8-1d4b-4302-84e4-7e30860f1864" (UID: "5f6065c8-1d4b-4302-84e4-7e30860f1864"). InnerVolumeSpecName "kube-api-access-ldqgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.404503 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f6065c8-1d4b-4302-84e4-7e30860f1864" (UID: "5f6065c8-1d4b-4302-84e4-7e30860f1864"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.405255 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory" (OuterVolumeSpecName: "inventory") pod "5f6065c8-1d4b-4302-84e4-7e30860f1864" (UID: "5f6065c8-1d4b-4302-84e4-7e30860f1864"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.453062 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldqgc\" (UniqueName: \"kubernetes.io/projected/5f6065c8-1d4b-4302-84e4-7e30860f1864-kube-api-access-ldqgc\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.453094 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.453106 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f6065c8-1d4b-4302-84e4-7e30860f1864-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.734286 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-jffww" event={"ID":"5f6065c8-1d4b-4302-84e4-7e30860f1864","Type":"ContainerDied","Data":"9bd0b1f07c094acd9a36ca3424c1716e98d66107fcd148b01029a92a39bbd2a4"} Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.734865 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bd0b1f07c094acd9a36ca3424c1716e98d66107fcd148b01029a92a39bbd2a4" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.734343 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-jffww" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.825859 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dz5sw"] Oct 01 09:34:16 crc kubenswrapper[4837]: E1001 09:34:16.826814 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6065c8-1d4b-4302-84e4-7e30860f1864" containerName="run-os-openstack-openstack-networker" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.826947 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6065c8-1d4b-4302-84e4-7e30860f1864" containerName="run-os-openstack-openstack-networker" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.827393 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6065c8-1d4b-4302-84e4-7e30860f1864" containerName="run-os-openstack-openstack-networker" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.828784 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.832937 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.836361 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.841243 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dz5sw"] Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.968988 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.969049 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:16 crc kubenswrapper[4837]: I1001 09:34:16.969171 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9jr4\" (UniqueName: \"kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.071604 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9jr4\" (UniqueName: \"kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.071746 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.071786 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.076498 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.082654 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.108098 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9jr4\" (UniqueName: \"kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4\") pod \"reboot-os-openstack-openstack-networker-dz5sw\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.161145 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:17 crc kubenswrapper[4837]: I1001 09:34:17.791524 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-dz5sw"] Oct 01 09:34:17 crc kubenswrapper[4837]: W1001 09:34:17.795504 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6106757_a810_46cd_8b2c_9e455dc9f95b.slice/crio-d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281 WatchSource:0}: Error finding container d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281: Status 404 returned error can't find the container with id d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281 Oct 01 09:34:18 crc kubenswrapper[4837]: I1001 09:34:18.776081 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" event={"ID":"c6106757-a810-46cd-8b2c-9e455dc9f95b","Type":"ContainerStarted","Data":"4f71f1b089f5bfa96e60f800161e0fd58d5a742faad84000870aee0be92f24d3"} Oct 01 09:34:18 crc kubenswrapper[4837]: I1001 09:34:18.776516 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" event={"ID":"c6106757-a810-46cd-8b2c-9e455dc9f95b","Type":"ContainerStarted","Data":"d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281"} Oct 01 09:34:18 crc kubenswrapper[4837]: I1001 09:34:18.795172 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" podStartSLOduration=2.213859938 podStartE2EDuration="2.795153776s" podCreationTimestamp="2025-10-01 09:34:16 +0000 UTC" firstStartedPulling="2025-10-01 09:34:17.798273371 +0000 UTC m=+8914.639880826" lastFinishedPulling="2025-10-01 09:34:18.379567179 +0000 UTC m=+8915.221174664" observedRunningTime="2025-10-01 09:34:18.791928337 +0000 UTC m=+8915.633535792" watchObservedRunningTime="2025-10-01 09:34:18.795153776 +0000 UTC m=+8915.636761251" Oct 01 09:34:22 crc kubenswrapper[4837]: I1001 09:34:22.827373 4837 generic.go:334] "Generic (PLEG): container finished" podID="587efc59-4844-4bb9-8467-5c39b985fab0" containerID="5f3094440856cf18259d9ed26882868f15fe55a0400c48d43ca590a193a8ed1f" exitCode=0 Oct 01 09:34:22 crc kubenswrapper[4837]: I1001 09:34:22.827489 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" event={"ID":"587efc59-4844-4bb9-8467-5c39b985fab0","Type":"ContainerDied","Data":"5f3094440856cf18259d9ed26882868f15fe55a0400c48d43ca590a193a8ed1f"} Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.079463 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.079551 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.079617 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.082671 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.082832 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217" gracePeriod=600 Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.865723 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217" exitCode=0 Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.866970 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217"} Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.867018 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172"} Oct 01 09:34:23 crc kubenswrapper[4837]: I1001 09:34:23.867050 4837 scope.go:117] "RemoveContainer" containerID="fb45ec0cc9679570892ec7469d907f5b10a5df8353f508626dc48f05887b935b" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.355079 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.433918 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory\") pod \"587efc59-4844-4bb9-8467-5c39b985fab0\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.434274 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key\") pod \"587efc59-4844-4bb9-8467-5c39b985fab0\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.434509 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df7sv\" (UniqueName: \"kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv\") pod \"587efc59-4844-4bb9-8467-5c39b985fab0\" (UID: \"587efc59-4844-4bb9-8467-5c39b985fab0\") " Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.444855 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv" (OuterVolumeSpecName: "kube-api-access-df7sv") pod "587efc59-4844-4bb9-8467-5c39b985fab0" (UID: "587efc59-4844-4bb9-8467-5c39b985fab0"). InnerVolumeSpecName "kube-api-access-df7sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.465728 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory" (OuterVolumeSpecName: "inventory") pod "587efc59-4844-4bb9-8467-5c39b985fab0" (UID: "587efc59-4844-4bb9-8467-5c39b985fab0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.471152 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "587efc59-4844-4bb9-8467-5c39b985fab0" (UID: "587efc59-4844-4bb9-8467-5c39b985fab0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.537510 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df7sv\" (UniqueName: \"kubernetes.io/projected/587efc59-4844-4bb9-8467-5c39b985fab0-kube-api-access-df7sv\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.537540 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.537548 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/587efc59-4844-4bb9-8467-5c39b985fab0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.905867 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.905876 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gmdpl" event={"ID":"587efc59-4844-4bb9-8467-5c39b985fab0","Type":"ContainerDied","Data":"f3ba83d011e733d212a3aedbf62a790fd0095c9ea44168e898810746c58fdc10"} Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.905987 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3ba83d011e733d212a3aedbf62a790fd0095c9ea44168e898810746c58fdc10" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.993543 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-tlfzz"] Oct 01 09:34:24 crc kubenswrapper[4837]: E1001 09:34:24.993996 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587efc59-4844-4bb9-8467-5c39b985fab0" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.994014 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="587efc59-4844-4bb9-8467-5c39b985fab0" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.994197 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="587efc59-4844-4bb9-8467-5c39b985fab0" containerName="configure-os-openstack-openstack-cell1" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.994947 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:24 crc kubenswrapper[4837]: I1001 09:34:24.997998 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.005683 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.011289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-tlfzz"] Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.151216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.151313 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.151580 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.151739 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.152067 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmls\" (UniqueName: \"kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.253991 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.254058 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.254179 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmls\" (UniqueName: \"kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.254232 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.254329 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.261921 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.262681 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.263756 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.270663 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.272379 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmls\" (UniqueName: \"kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls\") pod \"ssh-known-hosts-openstack-tlfzz\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.337833 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:25 crc kubenswrapper[4837]: I1001 09:34:25.947576 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-tlfzz"] Oct 01 09:34:25 crc kubenswrapper[4837]: W1001 09:34:25.948877 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod455fb732_e3bd_47a1_9428_b6a903332b76.slice/crio-c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab WatchSource:0}: Error finding container c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab: Status 404 returned error can't find the container with id c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab Oct 01 09:34:26 crc kubenswrapper[4837]: I1001 09:34:26.931289 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tlfzz" event={"ID":"455fb732-e3bd-47a1-9428-b6a903332b76","Type":"ContainerStarted","Data":"c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab"} Oct 01 09:34:27 crc kubenswrapper[4837]: I1001 09:34:27.951233 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tlfzz" event={"ID":"455fb732-e3bd-47a1-9428-b6a903332b76","Type":"ContainerStarted","Data":"70bbeab2e0fa9556be6ded5c7381e8857c228a10543a67a5c14fb97fe85c084d"} Oct 01 09:34:27 crc kubenswrapper[4837]: I1001 09:34:27.972041 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-tlfzz" podStartSLOduration=3.271019399 podStartE2EDuration="3.972021462s" podCreationTimestamp="2025-10-01 09:34:24 +0000 UTC" firstStartedPulling="2025-10-01 09:34:25.951663788 +0000 UTC m=+8922.793271253" lastFinishedPulling="2025-10-01 09:34:26.652665851 +0000 UTC m=+8923.494273316" observedRunningTime="2025-10-01 09:34:27.969532701 +0000 UTC m=+8924.811140176" watchObservedRunningTime="2025-10-01 09:34:27.972021462 +0000 UTC m=+8924.813628927" Oct 01 09:34:32 crc kubenswrapper[4837]: I1001 09:34:32.998941 4837 generic.go:334] "Generic (PLEG): container finished" podID="c6106757-a810-46cd-8b2c-9e455dc9f95b" containerID="4f71f1b089f5bfa96e60f800161e0fd58d5a742faad84000870aee0be92f24d3" exitCode=0 Oct 01 09:34:32 crc kubenswrapper[4837]: I1001 09:34:32.999037 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" event={"ID":"c6106757-a810-46cd-8b2c-9e455dc9f95b","Type":"ContainerDied","Data":"4f71f1b089f5bfa96e60f800161e0fd58d5a742faad84000870aee0be92f24d3"} Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.518964 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.675727 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9jr4\" (UniqueName: \"kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4\") pod \"c6106757-a810-46cd-8b2c-9e455dc9f95b\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.675949 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key\") pod \"c6106757-a810-46cd-8b2c-9e455dc9f95b\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.676009 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory\") pod \"c6106757-a810-46cd-8b2c-9e455dc9f95b\" (UID: \"c6106757-a810-46cd-8b2c-9e455dc9f95b\") " Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.683603 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4" (OuterVolumeSpecName: "kube-api-access-t9jr4") pod "c6106757-a810-46cd-8b2c-9e455dc9f95b" (UID: "c6106757-a810-46cd-8b2c-9e455dc9f95b"). InnerVolumeSpecName "kube-api-access-t9jr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.705017 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6106757-a810-46cd-8b2c-9e455dc9f95b" (UID: "c6106757-a810-46cd-8b2c-9e455dc9f95b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.735251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory" (OuterVolumeSpecName: "inventory") pod "c6106757-a810-46cd-8b2c-9e455dc9f95b" (UID: "c6106757-a810-46cd-8b2c-9e455dc9f95b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.779330 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.779380 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6106757-a810-46cd-8b2c-9e455dc9f95b-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:34 crc kubenswrapper[4837]: I1001 09:34:34.779402 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9jr4\" (UniqueName: \"kubernetes.io/projected/c6106757-a810-46cd-8b2c-9e455dc9f95b-kube-api-access-t9jr4\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.028447 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" event={"ID":"c6106757-a810-46cd-8b2c-9e455dc9f95b","Type":"ContainerDied","Data":"d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281"} Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.028945 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99b93b3572ef0704bc30c5f327f0d06f0108091ac3e3994a85a29d49f9f4281" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.029026 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-dz5sw" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.125353 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-networker-tvk69"] Oct 01 09:34:35 crc kubenswrapper[4837]: E1001 09:34:35.125841 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6106757-a810-46cd-8b2c-9e455dc9f95b" containerName="reboot-os-openstack-openstack-networker" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.125860 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6106757-a810-46cd-8b2c-9e455dc9f95b" containerName="reboot-os-openstack-openstack-networker" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.126112 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6106757-a810-46cd-8b2c-9e455dc9f95b" containerName="reboot-os-openstack-openstack-networker" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.126947 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.129720 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.130051 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-neutron-metadata-default-certs-0" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.130321 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-ovn-default-certs-0" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.154848 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-tvk69"] Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290532 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs5nw\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290578 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290637 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290675 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290770 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.290797 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.291000 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.291179 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.392956 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393047 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393093 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs5nw\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393147 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393172 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393238 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.393256 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.397831 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.401608 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.402524 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.402731 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.410418 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.421406 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.422115 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs5nw\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.422493 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-tvk69\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.443391 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:35 crc kubenswrapper[4837]: I1001 09:34:35.812136 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-tvk69"] Oct 01 09:34:36 crc kubenswrapper[4837]: I1001 09:34:36.062642 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-tvk69" event={"ID":"5d158e44-e94a-46e6-8fd5-82e31f63aff0","Type":"ContainerStarted","Data":"9347ac18d9f974ad5bfc31b71284a301b6eae39d76e8982b1a1d8c1f85e97ab9"} Oct 01 09:34:37 crc kubenswrapper[4837]: I1001 09:34:37.102396 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-tvk69" event={"ID":"5d158e44-e94a-46e6-8fd5-82e31f63aff0","Type":"ContainerStarted","Data":"392073762befac91d7c4efb3afee5003ca7dcb417e52eaddf5ceac4be4e39cb7"} Oct 01 09:34:37 crc kubenswrapper[4837]: I1001 09:34:37.155596 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-networker-tvk69" podStartSLOduration=1.724439071 podStartE2EDuration="2.155574982s" podCreationTimestamp="2025-10-01 09:34:35 +0000 UTC" firstStartedPulling="2025-10-01 09:34:35.822546055 +0000 UTC m=+8932.664153510" lastFinishedPulling="2025-10-01 09:34:36.253681956 +0000 UTC m=+8933.095289421" observedRunningTime="2025-10-01 09:34:37.143078615 +0000 UTC m=+8933.984686090" watchObservedRunningTime="2025-10-01 09:34:37.155574982 +0000 UTC m=+8933.997182447" Oct 01 09:34:42 crc kubenswrapper[4837]: I1001 09:34:42.150143 4837 generic.go:334] "Generic (PLEG): container finished" podID="455fb732-e3bd-47a1-9428-b6a903332b76" containerID="70bbeab2e0fa9556be6ded5c7381e8857c228a10543a67a5c14fb97fe85c084d" exitCode=0 Oct 01 09:34:42 crc kubenswrapper[4837]: I1001 09:34:42.150255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tlfzz" event={"ID":"455fb732-e3bd-47a1-9428-b6a903332b76","Type":"ContainerDied","Data":"70bbeab2e0fa9556be6ded5c7381e8857c228a10543a67a5c14fb97fe85c084d"} Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.686888 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.778201 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1\") pod \"455fb732-e3bd-47a1-9428-b6a903332b76\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.778312 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1\") pod \"455fb732-e3bd-47a1-9428-b6a903332b76\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.778405 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmls\" (UniqueName: \"kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls\") pod \"455fb732-e3bd-47a1-9428-b6a903332b76\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.778529 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0\") pod \"455fb732-e3bd-47a1-9428-b6a903332b76\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.778561 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker\") pod \"455fb732-e3bd-47a1-9428-b6a903332b76\" (UID: \"455fb732-e3bd-47a1-9428-b6a903332b76\") " Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.786070 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls" (OuterVolumeSpecName: "kube-api-access-hjmls") pod "455fb732-e3bd-47a1-9428-b6a903332b76" (UID: "455fb732-e3bd-47a1-9428-b6a903332b76"). InnerVolumeSpecName "kube-api-access-hjmls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.812860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "455fb732-e3bd-47a1-9428-b6a903332b76" (UID: "455fb732-e3bd-47a1-9428-b6a903332b76"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.824829 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "455fb732-e3bd-47a1-9428-b6a903332b76" (UID: "455fb732-e3bd-47a1-9428-b6a903332b76"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.844488 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1" (OuterVolumeSpecName: "inventory-1") pod "455fb732-e3bd-47a1-9428-b6a903332b76" (UID: "455fb732-e3bd-47a1-9428-b6a903332b76"). InnerVolumeSpecName "inventory-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.854430 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "455fb732-e3bd-47a1-9428-b6a903332b76" (UID: "455fb732-e3bd-47a1-9428-b6a903332b76"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.880461 4837 reconciler_common.go:293] "Volume detached for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-1\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.880722 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmls\" (UniqueName: \"kubernetes.io/projected/455fb732-e3bd-47a1-9428-b6a903332b76-kube-api-access-hjmls\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.880815 4837 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.880889 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:43 crc kubenswrapper[4837]: I1001 09:34:43.880968 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/455fb732-e3bd-47a1-9428-b6a903332b76-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.207462 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-tlfzz" event={"ID":"455fb732-e3bd-47a1-9428-b6a903332b76","Type":"ContainerDied","Data":"c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab"} Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.207757 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5410d461a58f0d74afce26ff2a6669e045ea6843382f3edf2dc416d042986ab" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.207545 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-tlfzz" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.251251 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-swmd5"] Oct 01 09:34:44 crc kubenswrapper[4837]: E1001 09:34:44.252081 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455fb732-e3bd-47a1-9428-b6a903332b76" containerName="ssh-known-hosts-openstack" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.252175 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="455fb732-e3bd-47a1-9428-b6a903332b76" containerName="ssh-known-hosts-openstack" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.252552 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="455fb732-e3bd-47a1-9428-b6a903332b76" containerName="ssh-known-hosts-openstack" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.253563 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.259273 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.263454 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-swmd5"] Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.264401 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.327801 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.328286 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fxwz\" (UniqueName: \"kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.328363 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.429958 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fxwz\" (UniqueName: \"kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.430017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.430206 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.437239 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.451769 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.453892 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fxwz\" (UniqueName: \"kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz\") pod \"run-os-openstack-openstack-cell1-swmd5\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.575332 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:44 crc kubenswrapper[4837]: I1001 09:34:44.960430 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-swmd5"] Oct 01 09:34:44 crc kubenswrapper[4837]: W1001 09:34:44.964615 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08a6e6cb_9507_4d0e_b23a_5816180342a8.slice/crio-ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1 WatchSource:0}: Error finding container ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1: Status 404 returned error can't find the container with id ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1 Oct 01 09:34:45 crc kubenswrapper[4837]: I1001 09:34:45.221352 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-swmd5" event={"ID":"08a6e6cb-9507-4d0e-b23a-5816180342a8","Type":"ContainerStarted","Data":"ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1"} Oct 01 09:34:46 crc kubenswrapper[4837]: I1001 09:34:46.234223 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-swmd5" event={"ID":"08a6e6cb-9507-4d0e-b23a-5816180342a8","Type":"ContainerStarted","Data":"546a80da15cdeeacd5affcc3dc7fd553566e2f48b810e33aacd3a0a22d0a9ba0"} Oct 01 09:34:46 crc kubenswrapper[4837]: I1001 09:34:46.265760 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-swmd5" podStartSLOduration=1.824528706 podStartE2EDuration="2.265728685s" podCreationTimestamp="2025-10-01 09:34:44 +0000 UTC" firstStartedPulling="2025-10-01 09:34:44.969073462 +0000 UTC m=+8941.810680927" lastFinishedPulling="2025-10-01 09:34:45.410273411 +0000 UTC m=+8942.251880906" observedRunningTime="2025-10-01 09:34:46.251779211 +0000 UTC m=+8943.093386696" watchObservedRunningTime="2025-10-01 09:34:46.265728685 +0000 UTC m=+8943.107336190" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.069007 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.072253 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.082060 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.235714 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phvl6\" (UniqueName: \"kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.236093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.236152 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.338071 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.338518 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.338593 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.338846 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.339030 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phvl6\" (UniqueName: \"kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.362801 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phvl6\" (UniqueName: \"kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6\") pod \"community-operators-rjgwb\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:54 crc kubenswrapper[4837]: I1001 09:34:54.391734 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:34:55 crc kubenswrapper[4837]: I1001 09:34:55.076113 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:34:55 crc kubenswrapper[4837]: I1001 09:34:55.363965 4837 generic.go:334] "Generic (PLEG): container finished" podID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerID="326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160" exitCode=0 Oct 01 09:34:55 crc kubenswrapper[4837]: I1001 09:34:55.364004 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerDied","Data":"326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160"} Oct 01 09:34:55 crc kubenswrapper[4837]: I1001 09:34:55.364030 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerStarted","Data":"8c0c15be6f12c11a13fffffd5eda0a2ad80c3d5008c0f5dcc10dcfabb3fce374"} Oct 01 09:34:56 crc kubenswrapper[4837]: I1001 09:34:56.379605 4837 generic.go:334] "Generic (PLEG): container finished" podID="08a6e6cb-9507-4d0e-b23a-5816180342a8" containerID="546a80da15cdeeacd5affcc3dc7fd553566e2f48b810e33aacd3a0a22d0a9ba0" exitCode=0 Oct 01 09:34:56 crc kubenswrapper[4837]: I1001 09:34:56.379737 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-swmd5" event={"ID":"08a6e6cb-9507-4d0e-b23a-5816180342a8","Type":"ContainerDied","Data":"546a80da15cdeeacd5affcc3dc7fd553566e2f48b810e33aacd3a0a22d0a9ba0"} Oct 01 09:34:56 crc kubenswrapper[4837]: I1001 09:34:56.383032 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerStarted","Data":"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7"} Oct 01 09:34:57 crc kubenswrapper[4837]: I1001 09:34:57.400846 4837 generic.go:334] "Generic (PLEG): container finished" podID="5d158e44-e94a-46e6-8fd5-82e31f63aff0" containerID="392073762befac91d7c4efb3afee5003ca7dcb417e52eaddf5ceac4be4e39cb7" exitCode=0 Oct 01 09:34:57 crc kubenswrapper[4837]: I1001 09:34:57.400992 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-tvk69" event={"ID":"5d158e44-e94a-46e6-8fd5-82e31f63aff0","Type":"ContainerDied","Data":"392073762befac91d7c4efb3afee5003ca7dcb417e52eaddf5ceac4be4e39cb7"} Oct 01 09:34:57 crc kubenswrapper[4837]: I1001 09:34:57.408069 4837 generic.go:334] "Generic (PLEG): container finished" podID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerID="d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7" exitCode=0 Oct 01 09:34:57 crc kubenswrapper[4837]: I1001 09:34:57.408197 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerDied","Data":"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7"} Oct 01 09:34:57 crc kubenswrapper[4837]: I1001 09:34:57.929360 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.017642 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fxwz\" (UniqueName: \"kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz\") pod \"08a6e6cb-9507-4d0e-b23a-5816180342a8\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.017731 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory\") pod \"08a6e6cb-9507-4d0e-b23a-5816180342a8\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.017893 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key\") pod \"08a6e6cb-9507-4d0e-b23a-5816180342a8\" (UID: \"08a6e6cb-9507-4d0e-b23a-5816180342a8\") " Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.022870 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz" (OuterVolumeSpecName: "kube-api-access-6fxwz") pod "08a6e6cb-9507-4d0e-b23a-5816180342a8" (UID: "08a6e6cb-9507-4d0e-b23a-5816180342a8"). InnerVolumeSpecName "kube-api-access-6fxwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.052871 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory" (OuterVolumeSpecName: "inventory") pod "08a6e6cb-9507-4d0e-b23a-5816180342a8" (UID: "08a6e6cb-9507-4d0e-b23a-5816180342a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.053305 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08a6e6cb-9507-4d0e-b23a-5816180342a8" (UID: "08a6e6cb-9507-4d0e-b23a-5816180342a8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.120791 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.120824 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fxwz\" (UniqueName: \"kubernetes.io/projected/08a6e6cb-9507-4d0e-b23a-5816180342a8-kube-api-access-6fxwz\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.120835 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08a6e6cb-9507-4d0e-b23a-5816180342a8-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.419654 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-swmd5" event={"ID":"08a6e6cb-9507-4d0e-b23a-5816180342a8","Type":"ContainerDied","Data":"ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1"} Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.420064 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffe40cb7650997ac3340539fac3a7e340ea4a9f6b2cb4b74c85635d94e61daa1" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.419667 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-swmd5" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.422999 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerStarted","Data":"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946"} Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.472553 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjgwb" podStartSLOduration=1.927514885 podStartE2EDuration="4.472530913s" podCreationTimestamp="2025-10-01 09:34:54 +0000 UTC" firstStartedPulling="2025-10-01 09:34:55.366413835 +0000 UTC m=+8952.208021290" lastFinishedPulling="2025-10-01 09:34:57.911429863 +0000 UTC m=+8954.753037318" observedRunningTime="2025-10-01 09:34:58.456319154 +0000 UTC m=+8955.297926639" watchObservedRunningTime="2025-10-01 09:34:58.472530913 +0000 UTC m=+8955.314138378" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.611838 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-n94sg"] Oct 01 09:34:58 crc kubenswrapper[4837]: E1001 09:34:58.612452 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a6e6cb-9507-4d0e-b23a-5816180342a8" containerName="run-os-openstack-openstack-cell1" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.612479 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a6e6cb-9507-4d0e-b23a-5816180342a8" containerName="run-os-openstack-openstack-cell1" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.612778 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a6e6cb-9507-4d0e-b23a-5816180342a8" containerName="run-os-openstack-openstack-cell1" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.613752 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.616926 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.617822 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.639486 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-n94sg"] Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.732282 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xmmq\" (UniqueName: \"kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.732875 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.733016 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.834754 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.834914 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.834963 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xmmq\" (UniqueName: \"kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.841283 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.841332 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.853265 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xmmq\" (UniqueName: \"kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq\") pod \"reboot-os-openstack-openstack-cell1-n94sg\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.926224 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:58 crc kubenswrapper[4837]: I1001 09:34:58.940031 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038137 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs5nw\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038278 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038301 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038376 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038475 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038504 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038546 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.038564 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0\") pod \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\" (UID: \"5d158e44-e94a-46e6-8fd5-82e31f63aff0\") " Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.041962 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.043560 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw" (OuterVolumeSpecName: "kube-api-access-gs5nw") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "kube-api-access-gs5nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.044600 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-networker-ovn-default-certs-0") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "openstack-networker-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.044830 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.045096 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-networker-neutron-metadata-default-certs-0") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "openstack-networker-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.052106 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.076235 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory" (OuterVolumeSpecName: "inventory") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.079775 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d158e44-e94a-46e6-8fd5-82e31f63aff0" (UID: "5d158e44-e94a-46e6-8fd5-82e31f63aff0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.141973 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142007 4837 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142017 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142027 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142038 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs5nw\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-kube-api-access-gs5nw\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142050 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142059 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d158e44-e94a-46e6-8fd5-82e31f63aff0-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.142068 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5d158e44-e94a-46e6-8fd5-82e31f63aff0-openstack-networker-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.436605 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-tvk69" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.437194 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-tvk69" event={"ID":"5d158e44-e94a-46e6-8fd5-82e31f63aff0","Type":"ContainerDied","Data":"9347ac18d9f974ad5bfc31b71284a301b6eae39d76e8982b1a1d8c1f85e97ab9"} Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.437289 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9347ac18d9f974ad5bfc31b71284a301b6eae39d76e8982b1a1d8c1f85e97ab9" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.544424 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-n94sg"] Oct 01 09:34:59 crc kubenswrapper[4837]: W1001 09:34:59.548324 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2540bf2_f9ab_41c6_a5ca_bee9e886cabd.slice/crio-96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74 WatchSource:0}: Error finding container 96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74: Status 404 returned error can't find the container with id 96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74 Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.554456 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-networker-nvhcc"] Oct 01 09:34:59 crc kubenswrapper[4837]: E1001 09:34:59.554868 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d158e44-e94a-46e6-8fd5-82e31f63aff0" containerName="install-certs-openstack-openstack-networker" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.554885 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d158e44-e94a-46e6-8fd5-82e31f63aff0" containerName="install-certs-openstack-openstack-networker" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.555083 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d158e44-e94a-46e6-8fd5-82e31f63aff0" containerName="install-certs-openstack-openstack-networker" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.555735 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.557582 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.557672 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.557591 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.590841 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-nvhcc"] Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.653219 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.653520 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s685x\" (UniqueName: \"kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.653549 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.653585 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.653636 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.755328 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.755369 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s685x\" (UniqueName: \"kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.755398 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.755431 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.755484 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.756308 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.760618 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.760868 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.761602 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.781480 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s685x\" (UniqueName: \"kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x\") pod \"ovn-openstack-openstack-networker-nvhcc\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:34:59 crc kubenswrapper[4837]: I1001 09:34:59.871260 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:35:00 crc kubenswrapper[4837]: I1001 09:35:00.446309 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" event={"ID":"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd","Type":"ContainerStarted","Data":"bf48ded755d8e3721067bf5d70f6006342880a221b4610b8d693ad603443fb27"} Oct 01 09:35:00 crc kubenswrapper[4837]: I1001 09:35:00.446584 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" event={"ID":"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd","Type":"ContainerStarted","Data":"96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74"} Oct 01 09:35:00 crc kubenswrapper[4837]: I1001 09:35:00.468319 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" podStartSLOduration=1.851978533 podStartE2EDuration="2.468289801s" podCreationTimestamp="2025-10-01 09:34:58 +0000 UTC" firstStartedPulling="2025-10-01 09:34:59.55117707 +0000 UTC m=+8956.392784525" lastFinishedPulling="2025-10-01 09:35:00.167488338 +0000 UTC m=+8957.009095793" observedRunningTime="2025-10-01 09:35:00.460404027 +0000 UTC m=+8957.302011492" watchObservedRunningTime="2025-10-01 09:35:00.468289801 +0000 UTC m=+8957.309897286" Oct 01 09:35:00 crc kubenswrapper[4837]: I1001 09:35:00.547464 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-nvhcc"] Oct 01 09:35:01 crc kubenswrapper[4837]: I1001 09:35:01.458434 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-nvhcc" event={"ID":"820b0167-6b33-4527-a211-e1ccbcf19a65","Type":"ContainerStarted","Data":"d35269f8ae92386997323d687ac73169c6d50f2042ed40ddd7eec88f1c1a4670"} Oct 01 09:35:01 crc kubenswrapper[4837]: I1001 09:35:01.459881 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-nvhcc" event={"ID":"820b0167-6b33-4527-a211-e1ccbcf19a65","Type":"ContainerStarted","Data":"121e90829ec606a702a81dbe1eb41fb67f835bef4400f4826b6792ff57bb26a0"} Oct 01 09:35:01 crc kubenswrapper[4837]: I1001 09:35:01.491498 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-networker-nvhcc" podStartSLOduration=1.8955502069999999 podStartE2EDuration="2.491477153s" podCreationTimestamp="2025-10-01 09:34:59 +0000 UTC" firstStartedPulling="2025-10-01 09:35:00.543912063 +0000 UTC m=+8957.385519528" lastFinishedPulling="2025-10-01 09:35:01.139839019 +0000 UTC m=+8957.981446474" observedRunningTime="2025-10-01 09:35:01.476434453 +0000 UTC m=+8958.318041928" watchObservedRunningTime="2025-10-01 09:35:01.491477153 +0000 UTC m=+8958.333084608" Oct 01 09:35:04 crc kubenswrapper[4837]: I1001 09:35:04.392448 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:04 crc kubenswrapper[4837]: I1001 09:35:04.392849 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:04 crc kubenswrapper[4837]: I1001 09:35:04.464451 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:04 crc kubenswrapper[4837]: I1001 09:35:04.551808 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:04 crc kubenswrapper[4837]: I1001 09:35:04.702552 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:35:06 crc kubenswrapper[4837]: I1001 09:35:06.511386 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rjgwb" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="registry-server" containerID="cri-o://2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946" gracePeriod=2 Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.063985 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.228060 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content\") pod \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.228356 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities\") pod \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.228469 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phvl6\" (UniqueName: \"kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6\") pod \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\" (UID: \"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3\") " Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.229247 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities" (OuterVolumeSpecName: "utilities") pod "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" (UID: "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.235335 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6" (OuterVolumeSpecName: "kube-api-access-phvl6") pod "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" (UID: "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3"). InnerVolumeSpecName "kube-api-access-phvl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.287649 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" (UID: "3590c4c6-0864-4ad6-bdcd-42946c1bcbb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.331924 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.331961 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phvl6\" (UniqueName: \"kubernetes.io/projected/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-kube-api-access-phvl6\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.331971 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.528154 4837 generic.go:334] "Generic (PLEG): container finished" podID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerID="2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946" exitCode=0 Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.528219 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerDied","Data":"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946"} Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.528260 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjgwb" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.528287 4837 scope.go:117] "RemoveContainer" containerID="2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.528268 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjgwb" event={"ID":"3590c4c6-0864-4ad6-bdcd-42946c1bcbb3","Type":"ContainerDied","Data":"8c0c15be6f12c11a13fffffd5eda0a2ad80c3d5008c0f5dcc10dcfabb3fce374"} Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.560713 4837 scope.go:117] "RemoveContainer" containerID="d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.588820 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.602850 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rjgwb"] Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.607076 4837 scope.go:117] "RemoveContainer" containerID="326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.666933 4837 scope.go:117] "RemoveContainer" containerID="2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946" Oct 01 09:35:07 crc kubenswrapper[4837]: E1001 09:35:07.667423 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946\": container with ID starting with 2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946 not found: ID does not exist" containerID="2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.667452 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946"} err="failed to get container status \"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946\": rpc error: code = NotFound desc = could not find container \"2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946\": container with ID starting with 2df8730c87431f46285595aed0cd89729ceaa16a490c6ec92378cb2e1868f946 not found: ID does not exist" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.667472 4837 scope.go:117] "RemoveContainer" containerID="d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7" Oct 01 09:35:07 crc kubenswrapper[4837]: E1001 09:35:07.668007 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7\": container with ID starting with d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7 not found: ID does not exist" containerID="d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.668027 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7"} err="failed to get container status \"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7\": rpc error: code = NotFound desc = could not find container \"d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7\": container with ID starting with d805f569700548b10a6b43a0c1527b6e6ad33f3f85efffc5f68ee891052b93c7 not found: ID does not exist" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.668039 4837 scope.go:117] "RemoveContainer" containerID="326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160" Oct 01 09:35:07 crc kubenswrapper[4837]: E1001 09:35:07.668265 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160\": container with ID starting with 326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160 not found: ID does not exist" containerID="326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.668284 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160"} err="failed to get container status \"326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160\": rpc error: code = NotFound desc = could not find container \"326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160\": container with ID starting with 326c77734384727ff98dda5b540c6e44d6e57bc9310f280b803fed41b2be5160 not found: ID does not exist" Oct 01 09:35:07 crc kubenswrapper[4837]: I1001 09:35:07.831068 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" path="/var/lib/kubelet/pods/3590c4c6-0864-4ad6-bdcd-42946c1bcbb3/volumes" Oct 01 09:35:15 crc kubenswrapper[4837]: I1001 09:35:15.625327 4837 generic.go:334] "Generic (PLEG): container finished" podID="b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" containerID="bf48ded755d8e3721067bf5d70f6006342880a221b4610b8d693ad603443fb27" exitCode=0 Oct 01 09:35:15 crc kubenswrapper[4837]: I1001 09:35:15.625532 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" event={"ID":"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd","Type":"ContainerDied","Data":"bf48ded755d8e3721067bf5d70f6006342880a221b4610b8d693ad603443fb27"} Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.277880 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.382833 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory\") pod \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.383281 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xmmq\" (UniqueName: \"kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq\") pod \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.383445 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key\") pod \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\" (UID: \"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd\") " Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.388858 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq" (OuterVolumeSpecName: "kube-api-access-2xmmq") pod "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" (UID: "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd"). InnerVolumeSpecName "kube-api-access-2xmmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.412838 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" (UID: "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.420439 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory" (OuterVolumeSpecName: "inventory") pod "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" (UID: "b2540bf2-f9ab-41c6-a5ca-bee9e886cabd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.486905 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.487175 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xmmq\" (UniqueName: \"kubernetes.io/projected/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-kube-api-access-2xmmq\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.487187 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2540bf2-f9ab-41c6-a5ca-bee9e886cabd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.648031 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" event={"ID":"b2540bf2-f9ab-41c6-a5ca-bee9e886cabd","Type":"ContainerDied","Data":"96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74"} Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.648075 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96aef176c98aaba982c86c7ea7e937b60d28bd50267a1702ac9940635047ce74" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.648149 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-n94sg" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.892320 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-zvqxc"] Oct 01 09:35:17 crc kubenswrapper[4837]: E1001 09:35:17.892759 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" containerName="reboot-os-openstack-openstack-cell1" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.892778 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" containerName="reboot-os-openstack-openstack-cell1" Oct 01 09:35:17 crc kubenswrapper[4837]: E1001 09:35:17.892806 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="registry-server" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.892813 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="registry-server" Oct 01 09:35:17 crc kubenswrapper[4837]: E1001 09:35:17.892841 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="extract-content" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.892848 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="extract-content" Oct 01 09:35:17 crc kubenswrapper[4837]: E1001 09:35:17.892867 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="extract-utilities" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.892873 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="extract-utilities" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.893065 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3590c4c6-0864-4ad6-bdcd-42946c1bcbb3" containerName="registry-server" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.893086 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2540bf2-f9ab-41c6-a5ca-bee9e886cabd" containerName="reboot-os-openstack-openstack-cell1" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.893758 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.897906 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.897969 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.898127 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.898169 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.898425 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.908454 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Oct 01 09:35:17 crc kubenswrapper[4837]: I1001 09:35:17.918844 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-zvqxc"] Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.000799 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001155 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001235 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001334 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001427 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001514 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001600 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001672 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001771 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv94f\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.001993 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.002062 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.002191 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.002326 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.104935 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.104999 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105032 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105070 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105089 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105110 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv94f\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105157 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105209 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105294 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105335 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105373 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105449 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105474 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.105500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.111007 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.111287 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.111585 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.112004 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.112449 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.112454 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.113982 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.114069 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.114513 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.128913 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.129017 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.129257 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.129460 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.129870 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.132056 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv94f\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f\") pod \"install-certs-openstack-openstack-cell1-zvqxc\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:18 crc kubenswrapper[4837]: I1001 09:35:18.217754 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:35:19 crc kubenswrapper[4837]: I1001 09:35:19.476537 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-zvqxc"] Oct 01 09:35:19 crc kubenswrapper[4837]: I1001 09:35:19.671962 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" event={"ID":"03068cc4-7997-4279-98d1-1e6705b0bd75","Type":"ContainerStarted","Data":"c16d636320b70c8f76b1ebececc9326398f3a62b2dc94dc6676e36b58d1cdf86"} Oct 01 09:35:21 crc kubenswrapper[4837]: I1001 09:35:21.722897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" event={"ID":"03068cc4-7997-4279-98d1-1e6705b0bd75","Type":"ContainerStarted","Data":"3af448918c572b394635971a23b6ad9b00786b804c07c07ff81950eab931952e"} Oct 01 09:35:21 crc kubenswrapper[4837]: I1001 09:35:21.755431 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" podStartSLOduration=3.884083893 podStartE2EDuration="4.755409597s" podCreationTimestamp="2025-10-01 09:35:17 +0000 UTC" firstStartedPulling="2025-10-01 09:35:19.481049531 +0000 UTC m=+8976.322656996" lastFinishedPulling="2025-10-01 09:35:20.352375235 +0000 UTC m=+8977.193982700" observedRunningTime="2025-10-01 09:35:21.749617373 +0000 UTC m=+8978.591224848" watchObservedRunningTime="2025-10-01 09:35:21.755409597 +0000 UTC m=+8978.597017052" Oct 01 09:36:07 crc kubenswrapper[4837]: I1001 09:36:07.244810 4837 generic.go:334] "Generic (PLEG): container finished" podID="03068cc4-7997-4279-98d1-1e6705b0bd75" containerID="3af448918c572b394635971a23b6ad9b00786b804c07c07ff81950eab931952e" exitCode=0 Oct 01 09:36:07 crc kubenswrapper[4837]: I1001 09:36:07.244959 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" event={"ID":"03068cc4-7997-4279-98d1-1e6705b0bd75","Type":"ContainerDied","Data":"3af448918c572b394635971a23b6ad9b00786b804c07c07ff81950eab931952e"} Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.838005 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906424 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906465 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906487 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906588 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906627 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906649 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906721 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906786 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906827 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906867 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv94f\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906887 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906915 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906933 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.906961 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle\") pod \"03068cc4-7997-4279-98d1-1e6705b0bd75\" (UID: \"03068cc4-7997-4279-98d1-1e6705b0bd75\") " Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.913227 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.913298 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.913869 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.914210 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.914933 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.916894 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.916989 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.917068 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.918251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.920717 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.920970 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f" (OuterVolumeSpecName: "kube-api-access-sv94f") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "kube-api-access-sv94f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.929917 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.932812 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.944569 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:08 crc kubenswrapper[4837]: I1001 09:36:08.947956 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory" (OuterVolumeSpecName: "inventory") pod "03068cc4-7997-4279-98d1-1e6705b0bd75" (UID: "03068cc4-7997-4279-98d1-1e6705b0bd75"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009714 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009743 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv94f\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-kube-api-access-sv94f\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009755 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009764 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009774 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009783 4837 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009792 4837 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009801 4837 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009810 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009819 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009830 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009838 4837 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009847 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009856 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03068cc4-7997-4279-98d1-1e6705b0bd75-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.009865 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/03068cc4-7997-4279-98d1-1e6705b0bd75-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.268393 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" event={"ID":"03068cc4-7997-4279-98d1-1e6705b0bd75","Type":"ContainerDied","Data":"c16d636320b70c8f76b1ebececc9326398f3a62b2dc94dc6676e36b58d1cdf86"} Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.268703 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c16d636320b70c8f76b1ebececc9326398f3a62b2dc94dc6676e36b58d1cdf86" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.268589 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-zvqxc" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.390882 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-2d8vv"] Oct 01 09:36:09 crc kubenswrapper[4837]: E1001 09:36:09.391569 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03068cc4-7997-4279-98d1-1e6705b0bd75" containerName="install-certs-openstack-openstack-cell1" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.391589 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="03068cc4-7997-4279-98d1-1e6705b0bd75" containerName="install-certs-openstack-openstack-cell1" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.391913 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="03068cc4-7997-4279-98d1-1e6705b0bd75" containerName="install-certs-openstack-openstack-cell1" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.392898 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.399239 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.399240 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.421898 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-2d8vv"] Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.436252 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.436325 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.436382 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gcns\" (UniqueName: \"kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.436429 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.436551 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.538149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.538208 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.538240 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.539106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gcns\" (UniqueName: \"kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.539150 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.539616 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.543910 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.545045 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.559465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.563113 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gcns\" (UniqueName: \"kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns\") pod \"ovn-openstack-openstack-cell1-2d8vv\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:09 crc kubenswrapper[4837]: I1001 09:36:09.728764 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:36:10 crc kubenswrapper[4837]: I1001 09:36:10.329078 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-2d8vv"] Oct 01 09:36:10 crc kubenswrapper[4837]: W1001 09:36:10.344704 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod907ebdee_b8d2_47ef_b934_615ec125f46b.slice/crio-1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85 WatchSource:0}: Error finding container 1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85: Status 404 returned error can't find the container with id 1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85 Oct 01 09:36:10 crc kubenswrapper[4837]: I1001 09:36:10.353284 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:36:11 crc kubenswrapper[4837]: I1001 09:36:11.295048 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" event={"ID":"907ebdee-b8d2-47ef-b934-615ec125f46b","Type":"ContainerStarted","Data":"c1cfbcaf8caead03bc55ab16a561aed39098d460bc57c69b79b01b64d9496d8c"} Oct 01 09:36:11 crc kubenswrapper[4837]: I1001 09:36:11.295479 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" event={"ID":"907ebdee-b8d2-47ef-b934-615ec125f46b","Type":"ContainerStarted","Data":"1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85"} Oct 01 09:36:11 crc kubenswrapper[4837]: I1001 09:36:11.319247 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" podStartSLOduration=1.7367667249999998 podStartE2EDuration="2.3192298s" podCreationTimestamp="2025-10-01 09:36:09 +0000 UTC" firstStartedPulling="2025-10-01 09:36:10.353086162 +0000 UTC m=+9027.194693617" lastFinishedPulling="2025-10-01 09:36:10.935549237 +0000 UTC m=+9027.777156692" observedRunningTime="2025-10-01 09:36:11.314297328 +0000 UTC m=+9028.155904793" watchObservedRunningTime="2025-10-01 09:36:11.3192298 +0000 UTC m=+9028.160837265" Oct 01 09:36:23 crc kubenswrapper[4837]: I1001 09:36:23.079561 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:36:23 crc kubenswrapper[4837]: I1001 09:36:23.080111 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:36:25 crc kubenswrapper[4837]: I1001 09:36:25.482083 4837 generic.go:334] "Generic (PLEG): container finished" podID="820b0167-6b33-4527-a211-e1ccbcf19a65" containerID="d35269f8ae92386997323d687ac73169c6d50f2042ed40ddd7eec88f1c1a4670" exitCode=0 Oct 01 09:36:25 crc kubenswrapper[4837]: I1001 09:36:25.482306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-nvhcc" event={"ID":"820b0167-6b33-4527-a211-e1ccbcf19a65","Type":"ContainerDied","Data":"d35269f8ae92386997323d687ac73169c6d50f2042ed40ddd7eec88f1c1a4670"} Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.060191 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.158313 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s685x\" (UniqueName: \"kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x\") pod \"820b0167-6b33-4527-a211-e1ccbcf19a65\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.158527 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0\") pod \"820b0167-6b33-4527-a211-e1ccbcf19a65\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.158593 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory\") pod \"820b0167-6b33-4527-a211-e1ccbcf19a65\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.158800 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle\") pod \"820b0167-6b33-4527-a211-e1ccbcf19a65\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.158900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key\") pod \"820b0167-6b33-4527-a211-e1ccbcf19a65\" (UID: \"820b0167-6b33-4527-a211-e1ccbcf19a65\") " Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.164317 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x" (OuterVolumeSpecName: "kube-api-access-s685x") pod "820b0167-6b33-4527-a211-e1ccbcf19a65" (UID: "820b0167-6b33-4527-a211-e1ccbcf19a65"). InnerVolumeSpecName "kube-api-access-s685x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.165055 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "820b0167-6b33-4527-a211-e1ccbcf19a65" (UID: "820b0167-6b33-4527-a211-e1ccbcf19a65"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.185290 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "820b0167-6b33-4527-a211-e1ccbcf19a65" (UID: "820b0167-6b33-4527-a211-e1ccbcf19a65"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.186007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory" (OuterVolumeSpecName: "inventory") pod "820b0167-6b33-4527-a211-e1ccbcf19a65" (UID: "820b0167-6b33-4527-a211-e1ccbcf19a65"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.188377 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "820b0167-6b33-4527-a211-e1ccbcf19a65" (UID: "820b0167-6b33-4527-a211-e1ccbcf19a65"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.262167 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.262211 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.262225 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s685x\" (UniqueName: \"kubernetes.io/projected/820b0167-6b33-4527-a211-e1ccbcf19a65-kube-api-access-s685x\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.262237 4837 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/820b0167-6b33-4527-a211-e1ccbcf19a65-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.262250 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/820b0167-6b33-4527-a211-e1ccbcf19a65-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.504201 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-nvhcc" event={"ID":"820b0167-6b33-4527-a211-e1ccbcf19a65","Type":"ContainerDied","Data":"121e90829ec606a702a81dbe1eb41fb67f835bef4400f4826b6792ff57bb26a0"} Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.504245 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="121e90829ec606a702a81dbe1eb41fb67f835bef4400f4826b6792ff57bb26a0" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.504315 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-nvhcc" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.601433 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-thkv6"] Oct 01 09:36:27 crc kubenswrapper[4837]: E1001 09:36:27.601860 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820b0167-6b33-4527-a211-e1ccbcf19a65" containerName="ovn-openstack-openstack-networker" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.601877 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="820b0167-6b33-4527-a211-e1ccbcf19a65" containerName="ovn-openstack-openstack-networker" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.602055 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="820b0167-6b33-4527-a211-e1ccbcf19a65" containerName="ovn-openstack-openstack-networker" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.602756 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.610399 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.610564 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.610565 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.610717 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-qvqqm" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.619819 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-thkv6"] Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.670430 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.670847 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.671011 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.671200 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.671376 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.671515 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkp5m\" (UniqueName: \"kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774443 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkp5m\" (UniqueName: \"kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774545 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774625 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774758 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774860 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:27 crc kubenswrapper[4837]: I1001 09:36:27.774984 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.651005 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.651243 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.651458 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.651750 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.652043 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.652487 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkp5m\" (UniqueName: \"kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m\") pod \"neutron-metadata-openstack-openstack-networker-thkv6\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:28 crc kubenswrapper[4837]: I1001 09:36:28.838418 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:36:29 crc kubenswrapper[4837]: W1001 09:36:29.411224 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffe682e2_2611_4971_a163_a2adfd4570d7.slice/crio-a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660 WatchSource:0}: Error finding container a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660: Status 404 returned error can't find the container with id a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660 Oct 01 09:36:29 crc kubenswrapper[4837]: I1001 09:36:29.417141 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-thkv6"] Oct 01 09:36:29 crc kubenswrapper[4837]: I1001 09:36:29.525504 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" event={"ID":"ffe682e2-2611-4971-a163-a2adfd4570d7","Type":"ContainerStarted","Data":"a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660"} Oct 01 09:36:30 crc kubenswrapper[4837]: I1001 09:36:30.539727 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" event={"ID":"ffe682e2-2611-4971-a163-a2adfd4570d7","Type":"ContainerStarted","Data":"7fb274f9e51a81d5a352da718d064b8a54316a5734dbc7b5b47ecda9ba835487"} Oct 01 09:36:30 crc kubenswrapper[4837]: I1001 09:36:30.568075 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" podStartSLOduration=3.114856636 podStartE2EDuration="3.56804586s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="2025-10-01 09:36:29.41407554 +0000 UTC m=+9046.255682995" lastFinishedPulling="2025-10-01 09:36:29.867264764 +0000 UTC m=+9046.708872219" observedRunningTime="2025-10-01 09:36:30.560025894 +0000 UTC m=+9047.401633389" watchObservedRunningTime="2025-10-01 09:36:30.56804586 +0000 UTC m=+9047.409653335" Oct 01 09:36:53 crc kubenswrapper[4837]: I1001 09:36:53.079000 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:36:53 crc kubenswrapper[4837]: I1001 09:36:53.079605 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:37:23 crc kubenswrapper[4837]: I1001 09:37:23.079845 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:37:23 crc kubenswrapper[4837]: I1001 09:37:23.081180 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:37:23 crc kubenswrapper[4837]: I1001 09:37:23.081258 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:37:23 crc kubenswrapper[4837]: I1001 09:37:23.082871 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:37:23 crc kubenswrapper[4837]: I1001 09:37:23.082967 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" gracePeriod=600 Oct 01 09:37:23 crc kubenswrapper[4837]: E1001 09:37:23.216919 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:37:24 crc kubenswrapper[4837]: I1001 09:37:24.198066 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" exitCode=0 Oct 01 09:37:24 crc kubenswrapper[4837]: I1001 09:37:24.198242 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172"} Oct 01 09:37:24 crc kubenswrapper[4837]: I1001 09:37:24.198426 4837 scope.go:117] "RemoveContainer" containerID="5577af7376dea803187a9b2515901c7b859272bd4c762f0d3e9c4e7601307217" Oct 01 09:37:24 crc kubenswrapper[4837]: I1001 09:37:24.199190 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:37:24 crc kubenswrapper[4837]: E1001 09:37:24.199512 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:37:29 crc kubenswrapper[4837]: I1001 09:37:29.270995 4837 generic.go:334] "Generic (PLEG): container finished" podID="907ebdee-b8d2-47ef-b934-615ec125f46b" containerID="c1cfbcaf8caead03bc55ab16a561aed39098d460bc57c69b79b01b64d9496d8c" exitCode=0 Oct 01 09:37:29 crc kubenswrapper[4837]: I1001 09:37:29.271095 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" event={"ID":"907ebdee-b8d2-47ef-b934-615ec125f46b","Type":"ContainerDied","Data":"c1cfbcaf8caead03bc55ab16a561aed39098d460bc57c69b79b01b64d9496d8c"} Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.292226 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.299135 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" event={"ID":"907ebdee-b8d2-47ef-b934-615ec125f46b","Type":"ContainerDied","Data":"1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85"} Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.299179 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d77451591eece4652ef4d3ac7d23e59c673df3bb1789e429220ebeb4b1c5e85" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.299190 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-2d8vv" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.389131 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle\") pod \"907ebdee-b8d2-47ef-b934-615ec125f46b\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.389396 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory\") pod \"907ebdee-b8d2-47ef-b934-615ec125f46b\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.389614 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key\") pod \"907ebdee-b8d2-47ef-b934-615ec125f46b\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.389672 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0\") pod \"907ebdee-b8d2-47ef-b934-615ec125f46b\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.389897 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gcns\" (UniqueName: \"kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns\") pod \"907ebdee-b8d2-47ef-b934-615ec125f46b\" (UID: \"907ebdee-b8d2-47ef-b934-615ec125f46b\") " Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.448205 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns" (OuterVolumeSpecName: "kube-api-access-6gcns") pod "907ebdee-b8d2-47ef-b934-615ec125f46b" (UID: "907ebdee-b8d2-47ef-b934-615ec125f46b"). InnerVolumeSpecName "kube-api-access-6gcns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.449423 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "907ebdee-b8d2-47ef-b934-615ec125f46b" (UID: "907ebdee-b8d2-47ef-b934-615ec125f46b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.449913 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "907ebdee-b8d2-47ef-b934-615ec125f46b" (UID: "907ebdee-b8d2-47ef-b934-615ec125f46b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.452942 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory" (OuterVolumeSpecName: "inventory") pod "907ebdee-b8d2-47ef-b934-615ec125f46b" (UID: "907ebdee-b8d2-47ef-b934-615ec125f46b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.455860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "907ebdee-b8d2-47ef-b934-615ec125f46b" (UID: "907ebdee-b8d2-47ef-b934-615ec125f46b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.493185 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gcns\" (UniqueName: \"kubernetes.io/projected/907ebdee-b8d2-47ef-b934-615ec125f46b-kube-api-access-6gcns\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.493226 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.493238 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.493261 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/907ebdee-b8d2-47ef-b934-615ec125f46b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:31 crc kubenswrapper[4837]: I1001 09:37:31.493271 4837 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/907ebdee-b8d2-47ef-b934-615ec125f46b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.436166 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xdctt"] Oct 01 09:37:32 crc kubenswrapper[4837]: E1001 09:37:32.437184 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907ebdee-b8d2-47ef-b934-615ec125f46b" containerName="ovn-openstack-openstack-cell1" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.437199 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="907ebdee-b8d2-47ef-b934-615ec125f46b" containerName="ovn-openstack-openstack-cell1" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.437452 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="907ebdee-b8d2-47ef-b934-615ec125f46b" containerName="ovn-openstack-openstack-cell1" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.438474 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.443621 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.443631 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.464413 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xdctt"] Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517002 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517053 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517081 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-557n2\" (UniqueName: \"kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517105 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517203 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.517274 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619617 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619661 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619681 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-557n2\" (UniqueName: \"kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619721 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619794 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.619865 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.627125 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.634443 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.634643 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.635174 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.635193 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.638829 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-557n2\" (UniqueName: \"kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2\") pod \"neutron-metadata-openstack-openstack-cell1-xdctt\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:32 crc kubenswrapper[4837]: I1001 09:37:32.758084 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:37:33 crc kubenswrapper[4837]: I1001 09:37:33.347840 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xdctt"] Oct 01 09:37:33 crc kubenswrapper[4837]: W1001 09:37:33.356486 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24d911bc_6b98_4fbf_a326_89be185e859e.slice/crio-43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c WatchSource:0}: Error finding container 43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c: Status 404 returned error can't find the container with id 43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c Oct 01 09:37:34 crc kubenswrapper[4837]: I1001 09:37:34.355150 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" event={"ID":"24d911bc-6b98-4fbf-a326-89be185e859e","Type":"ContainerStarted","Data":"43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c"} Oct 01 09:37:35 crc kubenswrapper[4837]: I1001 09:37:35.370474 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" event={"ID":"24d911bc-6b98-4fbf-a326-89be185e859e","Type":"ContainerStarted","Data":"d5e48f9a724213d64ce68b88838a33594bd91de5972614824a67598849b9ce72"} Oct 01 09:37:35 crc kubenswrapper[4837]: I1001 09:37:35.402064 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" podStartSLOduration=2.897137299 podStartE2EDuration="3.402044075s" podCreationTimestamp="2025-10-01 09:37:32 +0000 UTC" firstStartedPulling="2025-10-01 09:37:33.359743661 +0000 UTC m=+9110.201351116" lastFinishedPulling="2025-10-01 09:37:33.864650437 +0000 UTC m=+9110.706257892" observedRunningTime="2025-10-01 09:37:35.395810961 +0000 UTC m=+9112.237418456" watchObservedRunningTime="2025-10-01 09:37:35.402044075 +0000 UTC m=+9112.243651540" Oct 01 09:37:36 crc kubenswrapper[4837]: I1001 09:37:36.815880 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:37:36 crc kubenswrapper[4837]: E1001 09:37:36.816485 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:37:38 crc kubenswrapper[4837]: I1001 09:37:38.409755 4837 generic.go:334] "Generic (PLEG): container finished" podID="ffe682e2-2611-4971-a163-a2adfd4570d7" containerID="7fb274f9e51a81d5a352da718d064b8a54316a5734dbc7b5b47ecda9ba835487" exitCode=0 Oct 01 09:37:38 crc kubenswrapper[4837]: I1001 09:37:38.409880 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" event={"ID":"ffe682e2-2611-4971-a163-a2adfd4570d7","Type":"ContainerDied","Data":"7fb274f9e51a81d5a352da718d064b8a54316a5734dbc7b5b47ecda9ba835487"} Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.877808 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992221 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkp5m\" (UniqueName: \"kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992343 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992462 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992539 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992754 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.992874 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ffe682e2-2611-4971-a163-a2adfd4570d7\" (UID: \"ffe682e2-2611-4971-a163-a2adfd4570d7\") " Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.998978 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:39 crc kubenswrapper[4837]: I1001 09:37:39.998822 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m" (OuterVolumeSpecName: "kube-api-access-jkp5m") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "kube-api-access-jkp5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.032808 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.038453 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.045844 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory" (OuterVolumeSpecName: "inventory") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.046489 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ffe682e2-2611-4971-a163-a2adfd4570d7" (UID: "ffe682e2-2611-4971-a163-a2adfd4570d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095137 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095167 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095179 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkp5m\" (UniqueName: \"kubernetes.io/projected/ffe682e2-2611-4971-a163-a2adfd4570d7-kube-api-access-jkp5m\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095189 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095201 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.095209 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ffe682e2-2611-4971-a163-a2adfd4570d7-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.438639 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" event={"ID":"ffe682e2-2611-4971-a163-a2adfd4570d7","Type":"ContainerDied","Data":"a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660"} Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.438708 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a88a3a8b9a6a7ce2cf4cb57fc2ba52ca68702d6b396852bce7f37b83d29c4660" Oct 01 09:37:40 crc kubenswrapper[4837]: I1001 09:37:40.438774 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-thkv6" Oct 01 09:37:49 crc kubenswrapper[4837]: I1001 09:37:49.815750 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:37:49 crc kubenswrapper[4837]: E1001 09:37:49.816502 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:38:00 crc kubenswrapper[4837]: I1001 09:38:00.817417 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:38:00 crc kubenswrapper[4837]: E1001 09:38:00.818865 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:38:14 crc kubenswrapper[4837]: I1001 09:38:14.816006 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:38:14 crc kubenswrapper[4837]: E1001 09:38:14.816935 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:38:25 crc kubenswrapper[4837]: I1001 09:38:25.816125 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:38:25 crc kubenswrapper[4837]: E1001 09:38:25.820627 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:38:36 crc kubenswrapper[4837]: I1001 09:38:36.816166 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:38:36 crc kubenswrapper[4837]: E1001 09:38:36.816933 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:38:37 crc kubenswrapper[4837]: I1001 09:38:37.146602 4837 generic.go:334] "Generic (PLEG): container finished" podID="24d911bc-6b98-4fbf-a326-89be185e859e" containerID="d5e48f9a724213d64ce68b88838a33594bd91de5972614824a67598849b9ce72" exitCode=0 Oct 01 09:38:37 crc kubenswrapper[4837]: I1001 09:38:37.146762 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" event={"ID":"24d911bc-6b98-4fbf-a326-89be185e859e","Type":"ContainerDied","Data":"d5e48f9a724213d64ce68b88838a33594bd91de5972614824a67598849b9ce72"} Oct 01 09:38:38 crc kubenswrapper[4837]: I1001 09:38:38.916735 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038099 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-557n2\" (UniqueName: \"kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038203 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038257 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038347 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038487 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.038531 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle\") pod \"24d911bc-6b98-4fbf-a326-89be185e859e\" (UID: \"24d911bc-6b98-4fbf-a326-89be185e859e\") " Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.044977 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.045034 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2" (OuterVolumeSpecName: "kube-api-access-557n2") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "kube-api-access-557n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.069598 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.072130 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.078827 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.087035 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory" (OuterVolumeSpecName: "inventory") pod "24d911bc-6b98-4fbf-a326-89be185e859e" (UID: "24d911bc-6b98-4fbf-a326-89be185e859e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141237 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141277 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141288 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-557n2\" (UniqueName: \"kubernetes.io/projected/24d911bc-6b98-4fbf-a326-89be185e859e-kube-api-access-557n2\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141297 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141306 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.141314 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/24d911bc-6b98-4fbf-a326-89be185e859e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.168078 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" event={"ID":"24d911bc-6b98-4fbf-a326-89be185e859e","Type":"ContainerDied","Data":"43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c"} Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.168121 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43850142d8ebf94761b398439fe51a30649ad026523dd37d49f024e8c9990d3c" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.168170 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xdctt" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.284177 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7lfc2"] Oct 01 09:38:39 crc kubenswrapper[4837]: E1001 09:38:39.284652 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d911bc-6b98-4fbf-a326-89be185e859e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.284677 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d911bc-6b98-4fbf-a326-89be185e859e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 01 09:38:39 crc kubenswrapper[4837]: E1001 09:38:39.284732 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffe682e2-2611-4971-a163-a2adfd4570d7" containerName="neutron-metadata-openstack-openstack-networker" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.284746 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffe682e2-2611-4971-a163-a2adfd4570d7" containerName="neutron-metadata-openstack-openstack-networker" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.285022 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d911bc-6b98-4fbf-a326-89be185e859e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.285059 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffe682e2-2611-4971-a163-a2adfd4570d7" containerName="neutron-metadata-openstack-openstack-networker" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.285998 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.291635 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.292981 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.293201 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.293409 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.293588 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.303461 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7lfc2"] Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.350433 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.350611 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdgrd\" (UniqueName: \"kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.350766 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.350829 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.350857 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.453051 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.453123 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdgrd\" (UniqueName: \"kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.453213 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.453251 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.453275 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.458810 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.458816 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.458937 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.459567 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.472071 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdgrd\" (UniqueName: \"kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd\") pod \"libvirt-openstack-openstack-cell1-7lfc2\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:39 crc kubenswrapper[4837]: I1001 09:38:39.626091 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:38:40 crc kubenswrapper[4837]: I1001 09:38:40.174014 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-7lfc2"] Oct 01 09:38:41 crc kubenswrapper[4837]: I1001 09:38:41.189523 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" event={"ID":"442bd19f-fbab-4863-a8de-098783ec10e9","Type":"ContainerStarted","Data":"b1bfc90f23e8f97229db7bd1e5a9ad84d541fa4a1126d2d963ac6efd106341a1"} Oct 01 09:38:41 crc kubenswrapper[4837]: I1001 09:38:41.190537 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" event={"ID":"442bd19f-fbab-4863-a8de-098783ec10e9","Type":"ContainerStarted","Data":"4f0f2e6f573e8be8b268efd07390b701910eb023dcc96ec87e00698520f548ab"} Oct 01 09:38:41 crc kubenswrapper[4837]: I1001 09:38:41.227157 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" podStartSLOduration=1.8074922020000002 podStartE2EDuration="2.22711711s" podCreationTimestamp="2025-10-01 09:38:39 +0000 UTC" firstStartedPulling="2025-10-01 09:38:40.172649748 +0000 UTC m=+9177.014257203" lastFinishedPulling="2025-10-01 09:38:40.592274646 +0000 UTC m=+9177.433882111" observedRunningTime="2025-10-01 09:38:41.206427701 +0000 UTC m=+9178.048035166" watchObservedRunningTime="2025-10-01 09:38:41.22711711 +0000 UTC m=+9178.068724615" Oct 01 09:38:48 crc kubenswrapper[4837]: I1001 09:38:48.818260 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:38:48 crc kubenswrapper[4837]: E1001 09:38:48.819777 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.878509 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.885423 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.900637 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.994234 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.994371 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:01 crc kubenswrapper[4837]: I1001 09:39:01.994450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq2sr\" (UniqueName: \"kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.096477 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.096611 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.096717 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq2sr\" (UniqueName: \"kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.097223 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.097593 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.129648 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq2sr\" (UniqueName: \"kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr\") pod \"redhat-marketplace-ccj75\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.209411 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:02 crc kubenswrapper[4837]: I1001 09:39:02.816808 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:39:02 crc kubenswrapper[4837]: E1001 09:39:02.817438 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:03 crc kubenswrapper[4837]: I1001 09:39:03.341771 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:03 crc kubenswrapper[4837]: W1001 09:39:03.348212 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ad5ef8c_c3e2_42fd_9921_8ba60af545d1.slice/crio-7e5a8ddd2fbefa4f606267e3a8005032fb7db19fab90c391397cf9cac3a860b8 WatchSource:0}: Error finding container 7e5a8ddd2fbefa4f606267e3a8005032fb7db19fab90c391397cf9cac3a860b8: Status 404 returned error can't find the container with id 7e5a8ddd2fbefa4f606267e3a8005032fb7db19fab90c391397cf9cac3a860b8 Oct 01 09:39:03 crc kubenswrapper[4837]: I1001 09:39:03.449677 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerStarted","Data":"7e5a8ddd2fbefa4f606267e3a8005032fb7db19fab90c391397cf9cac3a860b8"} Oct 01 09:39:04 crc kubenswrapper[4837]: I1001 09:39:04.461069 4837 generic.go:334] "Generic (PLEG): container finished" podID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerID="ce823098151ec1324535d095c7e09b617baab5db9f0867312a8bc4d5b4b790f7" exitCode=0 Oct 01 09:39:04 crc kubenswrapper[4837]: I1001 09:39:04.461126 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerDied","Data":"ce823098151ec1324535d095c7e09b617baab5db9f0867312a8bc4d5b4b790f7"} Oct 01 09:39:06 crc kubenswrapper[4837]: I1001 09:39:06.485793 4837 generic.go:334] "Generic (PLEG): container finished" podID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerID="38dbc3c15d069f41545f1d77dea0248a377faeafc430fc8a49c3dcefa4b497c9" exitCode=0 Oct 01 09:39:06 crc kubenswrapper[4837]: I1001 09:39:06.485884 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerDied","Data":"38dbc3c15d069f41545f1d77dea0248a377faeafc430fc8a49c3dcefa4b497c9"} Oct 01 09:39:07 crc kubenswrapper[4837]: I1001 09:39:07.498009 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerStarted","Data":"93ef3466916475df219f604913c9fa974031a39d6b788e89d0fd7f1be29d659d"} Oct 01 09:39:07 crc kubenswrapper[4837]: I1001 09:39:07.527132 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ccj75" podStartSLOduration=4.033524378 podStartE2EDuration="6.527111648s" podCreationTimestamp="2025-10-01 09:39:01 +0000 UTC" firstStartedPulling="2025-10-01 09:39:04.464968995 +0000 UTC m=+9201.306576460" lastFinishedPulling="2025-10-01 09:39:06.958556265 +0000 UTC m=+9203.800163730" observedRunningTime="2025-10-01 09:39:07.521557202 +0000 UTC m=+9204.363164657" watchObservedRunningTime="2025-10-01 09:39:07.527111648 +0000 UTC m=+9204.368719103" Oct 01 09:39:12 crc kubenswrapper[4837]: I1001 09:39:12.211180 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:12 crc kubenswrapper[4837]: I1001 09:39:12.212793 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:12 crc kubenswrapper[4837]: I1001 09:39:12.277942 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:12 crc kubenswrapper[4837]: I1001 09:39:12.614140 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:12 crc kubenswrapper[4837]: I1001 09:39:12.668015 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:14 crc kubenswrapper[4837]: I1001 09:39:14.571786 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ccj75" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="registry-server" containerID="cri-o://93ef3466916475df219f604913c9fa974031a39d6b788e89d0fd7f1be29d659d" gracePeriod=2 Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.593570 4837 generic.go:334] "Generic (PLEG): container finished" podID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerID="93ef3466916475df219f604913c9fa974031a39d6b788e89d0fd7f1be29d659d" exitCode=0 Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.593622 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerDied","Data":"93ef3466916475df219f604913c9fa974031a39d6b788e89d0fd7f1be29d659d"} Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.771316 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.822809 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content\") pod \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.844562 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" (UID: "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.924325 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq2sr\" (UniqueName: \"kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr\") pod \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.924457 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities\") pod \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\" (UID: \"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1\") " Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.925075 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities" (OuterVolumeSpecName: "utilities") pod "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" (UID: "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.925210 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.925228 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:15 crc kubenswrapper[4837]: I1001 09:39:15.932121 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr" (OuterVolumeSpecName: "kube-api-access-hq2sr") pod "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" (UID: "0ad5ef8c-c3e2-42fd-9921-8ba60af545d1"). InnerVolumeSpecName "kube-api-access-hq2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.027300 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq2sr\" (UniqueName: \"kubernetes.io/projected/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1-kube-api-access-hq2sr\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.610434 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccj75" event={"ID":"0ad5ef8c-c3e2-42fd-9921-8ba60af545d1","Type":"ContainerDied","Data":"7e5a8ddd2fbefa4f606267e3a8005032fb7db19fab90c391397cf9cac3a860b8"} Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.610479 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccj75" Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.610510 4837 scope.go:117] "RemoveContainer" containerID="93ef3466916475df219f604913c9fa974031a39d6b788e89d0fd7f1be29d659d" Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.646241 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.646445 4837 scope.go:117] "RemoveContainer" containerID="38dbc3c15d069f41545f1d77dea0248a377faeafc430fc8a49c3dcefa4b497c9" Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.657684 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccj75"] Oct 01 09:39:16 crc kubenswrapper[4837]: I1001 09:39:16.816038 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:39:16 crc kubenswrapper[4837]: E1001 09:39:16.816528 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:17 crc kubenswrapper[4837]: I1001 09:39:17.467859 4837 scope.go:117] "RemoveContainer" containerID="ce823098151ec1324535d095c7e09b617baab5db9f0867312a8bc4d5b4b790f7" Oct 01 09:39:17 crc kubenswrapper[4837]: I1001 09:39:17.829656 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" path="/var/lib/kubelet/pods/0ad5ef8c-c3e2-42fd-9921-8ba60af545d1/volumes" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.624052 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:26 crc kubenswrapper[4837]: E1001 09:39:26.624867 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="extract-utilities" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.624881 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="extract-utilities" Oct 01 09:39:26 crc kubenswrapper[4837]: E1001 09:39:26.624893 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="extract-content" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.624899 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="extract-content" Oct 01 09:39:26 crc kubenswrapper[4837]: E1001 09:39:26.624922 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="registry-server" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.624928 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="registry-server" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.625118 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ad5ef8c-c3e2-42fd-9921-8ba60af545d1" containerName="registry-server" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.626488 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.648939 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.801578 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.801635 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7qcl\" (UniqueName: \"kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.801669 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.904140 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.904245 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7qcl\" (UniqueName: \"kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.904309 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.904728 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:26 crc kubenswrapper[4837]: I1001 09:39:26.904781 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:27 crc kubenswrapper[4837]: I1001 09:39:27.067621 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7qcl\" (UniqueName: \"kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl\") pod \"certified-operators-rhtz7\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:27 crc kubenswrapper[4837]: I1001 09:39:27.249538 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:27 crc kubenswrapper[4837]: I1001 09:39:27.765996 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:28 crc kubenswrapper[4837]: I1001 09:39:28.759632 4837 generic.go:334] "Generic (PLEG): container finished" podID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerID="8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba" exitCode=0 Oct 01 09:39:28 crc kubenswrapper[4837]: I1001 09:39:28.759757 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerDied","Data":"8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba"} Oct 01 09:39:28 crc kubenswrapper[4837]: I1001 09:39:28.760326 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerStarted","Data":"cce2ca380aaf61a1043ed80be45f74690fd7f41c4826f1311d10fe8c72fa0b28"} Oct 01 09:39:30 crc kubenswrapper[4837]: I1001 09:39:30.791016 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerStarted","Data":"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f"} Oct 01 09:39:31 crc kubenswrapper[4837]: I1001 09:39:31.802244 4837 generic.go:334] "Generic (PLEG): container finished" podID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerID="a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f" exitCode=0 Oct 01 09:39:31 crc kubenswrapper[4837]: I1001 09:39:31.802293 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerDied","Data":"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f"} Oct 01 09:39:31 crc kubenswrapper[4837]: I1001 09:39:31.821703 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:39:31 crc kubenswrapper[4837]: E1001 09:39:31.821943 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:32 crc kubenswrapper[4837]: I1001 09:39:32.814415 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerStarted","Data":"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20"} Oct 01 09:39:32 crc kubenswrapper[4837]: I1001 09:39:32.842790 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rhtz7" podStartSLOduration=3.356076299 podStartE2EDuration="6.842769772s" podCreationTimestamp="2025-10-01 09:39:26 +0000 UTC" firstStartedPulling="2025-10-01 09:39:28.76247687 +0000 UTC m=+9225.604084365" lastFinishedPulling="2025-10-01 09:39:32.249170383 +0000 UTC m=+9229.090777838" observedRunningTime="2025-10-01 09:39:32.832275793 +0000 UTC m=+9229.673883268" watchObservedRunningTime="2025-10-01 09:39:32.842769772 +0000 UTC m=+9229.684377247" Oct 01 09:39:37 crc kubenswrapper[4837]: I1001 09:39:37.250334 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:37 crc kubenswrapper[4837]: I1001 09:39:37.250942 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:37 crc kubenswrapper[4837]: I1001 09:39:37.312179 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:37 crc kubenswrapper[4837]: I1001 09:39:37.912340 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:37 crc kubenswrapper[4837]: I1001 09:39:37.970783 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:39 crc kubenswrapper[4837]: I1001 09:39:39.885665 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rhtz7" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="registry-server" containerID="cri-o://c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20" gracePeriod=2 Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.389894 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.451833 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7qcl\" (UniqueName: \"kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl\") pod \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.452060 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities\") pod \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.452201 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content\") pod \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\" (UID: \"7e538546-bab5-4c42-bcf5-d06a84bd4e97\") " Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.453187 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities" (OuterVolumeSpecName: "utilities") pod "7e538546-bab5-4c42-bcf5-d06a84bd4e97" (UID: "7e538546-bab5-4c42-bcf5-d06a84bd4e97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.457583 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl" (OuterVolumeSpecName: "kube-api-access-k7qcl") pod "7e538546-bab5-4c42-bcf5-d06a84bd4e97" (UID: "7e538546-bab5-4c42-bcf5-d06a84bd4e97"). InnerVolumeSpecName "kube-api-access-k7qcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.507363 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e538546-bab5-4c42-bcf5-d06a84bd4e97" (UID: "7e538546-bab5-4c42-bcf5-d06a84bd4e97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.555719 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7qcl\" (UniqueName: \"kubernetes.io/projected/7e538546-bab5-4c42-bcf5-d06a84bd4e97-kube-api-access-k7qcl\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.556081 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.556213 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e538546-bab5-4c42-bcf5-d06a84bd4e97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.897362 4837 generic.go:334] "Generic (PLEG): container finished" podID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerID="c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20" exitCode=0 Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.897424 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rhtz7" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.897409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerDied","Data":"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20"} Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.897843 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rhtz7" event={"ID":"7e538546-bab5-4c42-bcf5-d06a84bd4e97","Type":"ContainerDied","Data":"cce2ca380aaf61a1043ed80be45f74690fd7f41c4826f1311d10fe8c72fa0b28"} Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.897898 4837 scope.go:117] "RemoveContainer" containerID="c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.928764 4837 scope.go:117] "RemoveContainer" containerID="a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f" Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.944081 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:40 crc kubenswrapper[4837]: I1001 09:39:40.957218 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rhtz7"] Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.175512 4837 scope.go:117] "RemoveContainer" containerID="8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.214465 4837 scope.go:117] "RemoveContainer" containerID="c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20" Oct 01 09:39:41 crc kubenswrapper[4837]: E1001 09:39:41.214911 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20\": container with ID starting with c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20 not found: ID does not exist" containerID="c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.214940 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20"} err="failed to get container status \"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20\": rpc error: code = NotFound desc = could not find container \"c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20\": container with ID starting with c0c49497481f25bc053952ab06433daf731a2b48cee171aa8e2805cf5dd3ac20 not found: ID does not exist" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.214960 4837 scope.go:117] "RemoveContainer" containerID="a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f" Oct 01 09:39:41 crc kubenswrapper[4837]: E1001 09:39:41.215387 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f\": container with ID starting with a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f not found: ID does not exist" containerID="a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.215439 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f"} err="failed to get container status \"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f\": rpc error: code = NotFound desc = could not find container \"a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f\": container with ID starting with a29d86043b8f35240448e411443236fc7bf2e8ad3ad5922641d4a069d814ab0f not found: ID does not exist" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.215469 4837 scope.go:117] "RemoveContainer" containerID="8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba" Oct 01 09:39:41 crc kubenswrapper[4837]: E1001 09:39:41.215824 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba\": container with ID starting with 8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba not found: ID does not exist" containerID="8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.215852 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba"} err="failed to get container status \"8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba\": rpc error: code = NotFound desc = could not find container \"8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba\": container with ID starting with 8d7f8f7ca09e121e2c19cfbda83bbccbd9d8860867e9b6bc6ac107cdf7f0b5ba not found: ID does not exist" Oct 01 09:39:41 crc kubenswrapper[4837]: I1001 09:39:41.830085 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" path="/var/lib/kubelet/pods/7e538546-bab5-4c42-bcf5-d06a84bd4e97/volumes" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.816036 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:39:42 crc kubenswrapper[4837]: E1001 09:39:42.816575 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.974807 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:39:42 crc kubenswrapper[4837]: E1001 09:39:42.975440 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="extract-utilities" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.975460 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="extract-utilities" Oct 01 09:39:42 crc kubenswrapper[4837]: E1001 09:39:42.975473 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="registry-server" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.975480 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="registry-server" Oct 01 09:39:42 crc kubenswrapper[4837]: E1001 09:39:42.975525 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="extract-content" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.975534 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="extract-content" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.975792 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e538546-bab5-4c42-bcf5-d06a84bd4e97" containerName="registry-server" Oct 01 09:39:42 crc kubenswrapper[4837]: I1001 09:39:42.977589 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.005542 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.107726 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.108336 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktqcp\" (UniqueName: \"kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.108392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.210122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktqcp\" (UniqueName: \"kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.210236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.210346 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.210983 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.210991 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.232931 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktqcp\" (UniqueName: \"kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp\") pod \"redhat-operators-clvtw\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.299869 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.784742 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:39:43 crc kubenswrapper[4837]: I1001 09:39:43.937681 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerStarted","Data":"a0f7e56b3cc9a2da20dbf6f90e236180f2667d9451fedb88b28f9f8ecafd037d"} Oct 01 09:39:44 crc kubenswrapper[4837]: I1001 09:39:44.957514 4837 generic.go:334] "Generic (PLEG): container finished" podID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerID="83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494" exitCode=0 Oct 01 09:39:44 crc kubenswrapper[4837]: I1001 09:39:44.957656 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerDied","Data":"83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494"} Oct 01 09:39:46 crc kubenswrapper[4837]: I1001 09:39:46.978965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerStarted","Data":"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e"} Oct 01 09:39:51 crc kubenswrapper[4837]: I1001 09:39:51.039011 4837 generic.go:334] "Generic (PLEG): container finished" podID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerID="d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e" exitCode=0 Oct 01 09:39:51 crc kubenswrapper[4837]: I1001 09:39:51.039109 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerDied","Data":"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e"} Oct 01 09:39:53 crc kubenswrapper[4837]: I1001 09:39:53.064836 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerStarted","Data":"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16"} Oct 01 09:39:53 crc kubenswrapper[4837]: I1001 09:39:53.089728 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-clvtw" podStartSLOduration=4.374119677 podStartE2EDuration="11.089652024s" podCreationTimestamp="2025-10-01 09:39:42 +0000 UTC" firstStartedPulling="2025-10-01 09:39:44.96027686 +0000 UTC m=+9241.801884345" lastFinishedPulling="2025-10-01 09:39:51.675809197 +0000 UTC m=+9248.517416692" observedRunningTime="2025-10-01 09:39:53.086121558 +0000 UTC m=+9249.927729023" watchObservedRunningTime="2025-10-01 09:39:53.089652024 +0000 UTC m=+9249.931259519" Oct 01 09:39:53 crc kubenswrapper[4837]: I1001 09:39:53.300246 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:53 crc kubenswrapper[4837]: I1001 09:39:53.300454 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:39:53 crc kubenswrapper[4837]: I1001 09:39:53.846273 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:39:53 crc kubenswrapper[4837]: E1001 09:39:53.847393 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:39:54 crc kubenswrapper[4837]: I1001 09:39:54.500647 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-clvtw" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" probeResult="failure" output=< Oct 01 09:39:54 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:39:54 crc kubenswrapper[4837]: > Oct 01 09:40:04 crc kubenswrapper[4837]: I1001 09:40:04.809026 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-clvtw" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" probeResult="failure" output=< Oct 01 09:40:04 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:40:04 crc kubenswrapper[4837]: > Oct 01 09:40:08 crc kubenswrapper[4837]: I1001 09:40:08.816970 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:40:08 crc kubenswrapper[4837]: E1001 09:40:08.818040 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:40:13 crc kubenswrapper[4837]: I1001 09:40:13.372064 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:40:13 crc kubenswrapper[4837]: I1001 09:40:13.441406 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:40:14 crc kubenswrapper[4837]: I1001 09:40:14.188975 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.327835 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-clvtw" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" containerID="cri-o://7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16" gracePeriod=2 Oct 01 09:40:15 crc kubenswrapper[4837]: E1001 09:40:15.581747 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb37d229_111b_4e2a_b85c_6e6e7e82f940.slice/crio-7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb37d229_111b_4e2a_b85c_6e6e7e82f940.slice/crio-conmon-7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.834238 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.960455 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktqcp\" (UniqueName: \"kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp\") pod \"db37d229-111b-4e2a-b85c-6e6e7e82f940\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.960741 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content\") pod \"db37d229-111b-4e2a-b85c-6e6e7e82f940\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.960825 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities\") pod \"db37d229-111b-4e2a-b85c-6e6e7e82f940\" (UID: \"db37d229-111b-4e2a-b85c-6e6e7e82f940\") " Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.961490 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities" (OuterVolumeSpecName: "utilities") pod "db37d229-111b-4e2a-b85c-6e6e7e82f940" (UID: "db37d229-111b-4e2a-b85c-6e6e7e82f940"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:15 crc kubenswrapper[4837]: I1001 09:40:15.967601 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp" (OuterVolumeSpecName: "kube-api-access-ktqcp") pod "db37d229-111b-4e2a-b85c-6e6e7e82f940" (UID: "db37d229-111b-4e2a-b85c-6e6e7e82f940"). InnerVolumeSpecName "kube-api-access-ktqcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.063749 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktqcp\" (UniqueName: \"kubernetes.io/projected/db37d229-111b-4e2a-b85c-6e6e7e82f940-kube-api-access-ktqcp\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.063780 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.068949 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db37d229-111b-4e2a-b85c-6e6e7e82f940" (UID: "db37d229-111b-4e2a-b85c-6e6e7e82f940"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.166386 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db37d229-111b-4e2a-b85c-6e6e7e82f940-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.344895 4837 generic.go:334] "Generic (PLEG): container finished" podID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerID="7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16" exitCode=0 Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.344947 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerDied","Data":"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16"} Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.345007 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clvtw" event={"ID":"db37d229-111b-4e2a-b85c-6e6e7e82f940","Type":"ContainerDied","Data":"a0f7e56b3cc9a2da20dbf6f90e236180f2667d9451fedb88b28f9f8ecafd037d"} Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.345031 4837 scope.go:117] "RemoveContainer" containerID="7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.345045 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clvtw" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.378930 4837 scope.go:117] "RemoveContainer" containerID="d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.426570 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.448056 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-clvtw"] Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.449899 4837 scope.go:117] "RemoveContainer" containerID="83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.496389 4837 scope.go:117] "RemoveContainer" containerID="7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16" Oct 01 09:40:16 crc kubenswrapper[4837]: E1001 09:40:16.497030 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16\": container with ID starting with 7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16 not found: ID does not exist" containerID="7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.497099 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16"} err="failed to get container status \"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16\": rpc error: code = NotFound desc = could not find container \"7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16\": container with ID starting with 7fc38b960b37f752a950e40dd154540400ff0ab8463049669f6cd9e0fd383d16 not found: ID does not exist" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.497149 4837 scope.go:117] "RemoveContainer" containerID="d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e" Oct 01 09:40:16 crc kubenswrapper[4837]: E1001 09:40:16.497525 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e\": container with ID starting with d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e not found: ID does not exist" containerID="d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.497555 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e"} err="failed to get container status \"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e\": rpc error: code = NotFound desc = could not find container \"d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e\": container with ID starting with d190e3060b7553a5a4b6f95565ecc93ee019a409e58bf3c95c6b7a3a42bccf5e not found: ID does not exist" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.497577 4837 scope.go:117] "RemoveContainer" containerID="83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494" Oct 01 09:40:16 crc kubenswrapper[4837]: E1001 09:40:16.497857 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494\": container with ID starting with 83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494 not found: ID does not exist" containerID="83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494" Oct 01 09:40:16 crc kubenswrapper[4837]: I1001 09:40:16.497884 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494"} err="failed to get container status \"83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494\": rpc error: code = NotFound desc = could not find container \"83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494\": container with ID starting with 83172374f447a0f133740879ccdcf7083a1630bb30cd6dbed191c94e32287494 not found: ID does not exist" Oct 01 09:40:17 crc kubenswrapper[4837]: I1001 09:40:17.840165 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" path="/var/lib/kubelet/pods/db37d229-111b-4e2a-b85c-6e6e7e82f940/volumes" Oct 01 09:40:19 crc kubenswrapper[4837]: I1001 09:40:19.816201 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:40:19 crc kubenswrapper[4837]: E1001 09:40:19.817464 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:40:30 crc kubenswrapper[4837]: I1001 09:40:30.817096 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:40:30 crc kubenswrapper[4837]: E1001 09:40:30.818171 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:40:42 crc kubenswrapper[4837]: I1001 09:40:42.816558 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:40:42 crc kubenswrapper[4837]: E1001 09:40:42.817747 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:40:53 crc kubenswrapper[4837]: I1001 09:40:53.831031 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:40:53 crc kubenswrapper[4837]: E1001 09:40:53.832759 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:41:04 crc kubenswrapper[4837]: I1001 09:41:04.816859 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:41:04 crc kubenswrapper[4837]: E1001 09:41:04.817859 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:41:19 crc kubenswrapper[4837]: I1001 09:41:19.816901 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:41:19 crc kubenswrapper[4837]: E1001 09:41:19.817805 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:41:31 crc kubenswrapper[4837]: I1001 09:41:31.816984 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:41:31 crc kubenswrapper[4837]: E1001 09:41:31.819232 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:41:42 crc kubenswrapper[4837]: I1001 09:41:42.816277 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:41:42 crc kubenswrapper[4837]: E1001 09:41:42.817178 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:41:54 crc kubenswrapper[4837]: I1001 09:41:54.816745 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:41:54 crc kubenswrapper[4837]: E1001 09:41:54.817780 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:42:08 crc kubenswrapper[4837]: I1001 09:42:08.816274 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:42:08 crc kubenswrapper[4837]: E1001 09:42:08.817235 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:42:22 crc kubenswrapper[4837]: I1001 09:42:22.816917 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:42:22 crc kubenswrapper[4837]: E1001 09:42:22.818231 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:42:37 crc kubenswrapper[4837]: I1001 09:42:37.817009 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:42:38 crc kubenswrapper[4837]: I1001 09:42:38.248593 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7"} Oct 01 09:43:21 crc kubenswrapper[4837]: I1001 09:43:21.830598 4837 generic.go:334] "Generic (PLEG): container finished" podID="442bd19f-fbab-4863-a8de-098783ec10e9" containerID="b1bfc90f23e8f97229db7bd1e5a9ad84d541fa4a1126d2d963ac6efd106341a1" exitCode=0 Oct 01 09:43:21 crc kubenswrapper[4837]: I1001 09:43:21.844456 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" event={"ID":"442bd19f-fbab-4863-a8de-098783ec10e9","Type":"ContainerDied","Data":"b1bfc90f23e8f97229db7bd1e5a9ad84d541fa4a1126d2d963ac6efd106341a1"} Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.352497 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.495088 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key\") pod \"442bd19f-fbab-4863-a8de-098783ec10e9\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.495137 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory\") pod \"442bd19f-fbab-4863-a8de-098783ec10e9\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.495184 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0\") pod \"442bd19f-fbab-4863-a8de-098783ec10e9\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.495286 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle\") pod \"442bd19f-fbab-4863-a8de-098783ec10e9\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.495438 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdgrd\" (UniqueName: \"kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd\") pod \"442bd19f-fbab-4863-a8de-098783ec10e9\" (UID: \"442bd19f-fbab-4863-a8de-098783ec10e9\") " Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.503966 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "442bd19f-fbab-4863-a8de-098783ec10e9" (UID: "442bd19f-fbab-4863-a8de-098783ec10e9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.504520 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd" (OuterVolumeSpecName: "kube-api-access-kdgrd") pod "442bd19f-fbab-4863-a8de-098783ec10e9" (UID: "442bd19f-fbab-4863-a8de-098783ec10e9"). InnerVolumeSpecName "kube-api-access-kdgrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.553771 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "442bd19f-fbab-4863-a8de-098783ec10e9" (UID: "442bd19f-fbab-4863-a8de-098783ec10e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.554469 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory" (OuterVolumeSpecName: "inventory") pod "442bd19f-fbab-4863-a8de-098783ec10e9" (UID: "442bd19f-fbab-4863-a8de-098783ec10e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.566556 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "442bd19f-fbab-4863-a8de-098783ec10e9" (UID: "442bd19f-fbab-4863-a8de-098783ec10e9"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.605271 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.605313 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.605329 4837 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.605343 4837 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442bd19f-fbab-4863-a8de-098783ec10e9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.605355 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdgrd\" (UniqueName: \"kubernetes.io/projected/442bd19f-fbab-4863-a8de-098783ec10e9-kube-api-access-kdgrd\") on node \"crc\" DevicePath \"\"" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.866041 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" event={"ID":"442bd19f-fbab-4863-a8de-098783ec10e9","Type":"ContainerDied","Data":"4f0f2e6f573e8be8b268efd07390b701910eb023dcc96ec87e00698520f548ab"} Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.866105 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f0f2e6f573e8be8b268efd07390b701910eb023dcc96ec87e00698520f548ab" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.866150 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-7lfc2" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.965514 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-rjp2f"] Oct 01 09:43:23 crc kubenswrapper[4837]: E1001 09:43:23.966289 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966323 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" Oct 01 09:43:23 crc kubenswrapper[4837]: E1001 09:43:23.966390 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="extract-content" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966404 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="extract-content" Oct 01 09:43:23 crc kubenswrapper[4837]: E1001 09:43:23.966430 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442bd19f-fbab-4863-a8de-098783ec10e9" containerName="libvirt-openstack-openstack-cell1" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966445 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="442bd19f-fbab-4863-a8de-098783ec10e9" containerName="libvirt-openstack-openstack-cell1" Oct 01 09:43:23 crc kubenswrapper[4837]: E1001 09:43:23.966497 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="extract-utilities" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966510 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="extract-utilities" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966882 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="db37d229-111b-4e2a-b85c-6e6e7e82f940" containerName="registry-server" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.966942 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="442bd19f-fbab-4863-a8de-098783ec10e9" containerName="libvirt-openstack-openstack-cell1" Oct 01 09:43:23 crc kubenswrapper[4837]: I1001 09:43:23.968219 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.972488 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.972678 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.972852 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.972967 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.973080 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.973175 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.974541 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:23.975552 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-rjp2f"] Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.116259 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.116664 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bcqp\" (UniqueName: \"kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.116807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.116974 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.117102 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.117404 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.117495 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.117612 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.117746 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.220604 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.220743 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.220801 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.220861 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.220979 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.221025 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bcqp\" (UniqueName: \"kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.221113 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.221159 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.221212 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.221621 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.226042 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.226199 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.226615 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.229278 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.230353 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.230519 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.230742 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.243889 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bcqp\" (UniqueName: \"kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp\") pod \"nova-cell1-openstack-openstack-cell1-rjp2f\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.328681 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.937790 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:43:24 crc kubenswrapper[4837]: I1001 09:43:24.944860 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-rjp2f"] Oct 01 09:43:25 crc kubenswrapper[4837]: I1001 09:43:25.890630 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" event={"ID":"58f455af-4ff9-4beb-a59f-8ccfc4860c67","Type":"ContainerStarted","Data":"43c6aec58b562edfa55d55bce68960a9f3a2abc48ae525a70e6d4ab4adc0f428"} Oct 01 09:43:25 crc kubenswrapper[4837]: I1001 09:43:25.890670 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" event={"ID":"58f455af-4ff9-4beb-a59f-8ccfc4860c67","Type":"ContainerStarted","Data":"308eac78948aed843d877e7f6383d062aeecc5ac8b540212ad49a34fdcfd461a"} Oct 01 09:43:25 crc kubenswrapper[4837]: I1001 09:43:25.909870 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" podStartSLOduration=2.436837395 podStartE2EDuration="2.909849483s" podCreationTimestamp="2025-10-01 09:43:23 +0000 UTC" firstStartedPulling="2025-10-01 09:43:24.93754352 +0000 UTC m=+9461.779150985" lastFinishedPulling="2025-10-01 09:43:25.410555608 +0000 UTC m=+9462.252163073" observedRunningTime="2025-10-01 09:43:25.908834339 +0000 UTC m=+9462.750441794" watchObservedRunningTime="2025-10-01 09:43:25.909849483 +0000 UTC m=+9462.751456938" Oct 01 09:44:53 crc kubenswrapper[4837]: I1001 09:44:53.079216 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:44:53 crc kubenswrapper[4837]: I1001 09:44:53.080010 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.173756 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h"] Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.177457 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.180461 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.180925 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.185346 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h"] Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.292588 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.292817 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mwd\" (UniqueName: \"kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.292873 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.395163 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mwd\" (UniqueName: \"kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.395240 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.395290 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.396425 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.414491 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mwd\" (UniqueName: \"kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.414892 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume\") pod \"collect-profiles-29321865-vzb2h\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:00 crc kubenswrapper[4837]: I1001 09:45:00.502076 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:01 crc kubenswrapper[4837]: I1001 09:45:01.004049 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h"] Oct 01 09:45:01 crc kubenswrapper[4837]: I1001 09:45:01.158965 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" event={"ID":"5b0ead52-9298-441f-bca2-7fadf774448b","Type":"ContainerStarted","Data":"91e057aab1839a8d52ec6d73f14ee239a98c4a40d84dd303dd21969f016bd0fd"} Oct 01 09:45:02 crc kubenswrapper[4837]: I1001 09:45:02.176143 4837 generic.go:334] "Generic (PLEG): container finished" podID="5b0ead52-9298-441f-bca2-7fadf774448b" containerID="2847180ad0b9adc1a6d26ad7a34cff906738ccbb64a3cd6686c7889fb5e172c0" exitCode=0 Oct 01 09:45:02 crc kubenswrapper[4837]: I1001 09:45:02.176259 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" event={"ID":"5b0ead52-9298-441f-bca2-7fadf774448b","Type":"ContainerDied","Data":"2847180ad0b9adc1a6d26ad7a34cff906738ccbb64a3cd6686c7889fb5e172c0"} Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.674361 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.808763 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2mwd\" (UniqueName: \"kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd\") pod \"5b0ead52-9298-441f-bca2-7fadf774448b\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.809308 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume\") pod \"5b0ead52-9298-441f-bca2-7fadf774448b\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.809464 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume\") pod \"5b0ead52-9298-441f-bca2-7fadf774448b\" (UID: \"5b0ead52-9298-441f-bca2-7fadf774448b\") " Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.810056 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b0ead52-9298-441f-bca2-7fadf774448b" (UID: "5b0ead52-9298-441f-bca2-7fadf774448b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.817713 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd" (OuterVolumeSpecName: "kube-api-access-z2mwd") pod "5b0ead52-9298-441f-bca2-7fadf774448b" (UID: "5b0ead52-9298-441f-bca2-7fadf774448b"). InnerVolumeSpecName "kube-api-access-z2mwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.817807 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b0ead52-9298-441f-bca2-7fadf774448b" (UID: "5b0ead52-9298-441f-bca2-7fadf774448b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.913014 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2mwd\" (UniqueName: \"kubernetes.io/projected/5b0ead52-9298-441f-bca2-7fadf774448b-kube-api-access-z2mwd\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.913044 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b0ead52-9298-441f-bca2-7fadf774448b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:03 crc kubenswrapper[4837]: I1001 09:45:03.913053 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b0ead52-9298-441f-bca2-7fadf774448b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:04 crc kubenswrapper[4837]: I1001 09:45:04.201727 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" event={"ID":"5b0ead52-9298-441f-bca2-7fadf774448b","Type":"ContainerDied","Data":"91e057aab1839a8d52ec6d73f14ee239a98c4a40d84dd303dd21969f016bd0fd"} Oct 01 09:45:04 crc kubenswrapper[4837]: I1001 09:45:04.202157 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91e057aab1839a8d52ec6d73f14ee239a98c4a40d84dd303dd21969f016bd0fd" Oct 01 09:45:04 crc kubenswrapper[4837]: I1001 09:45:04.201799 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h" Oct 01 09:45:04 crc kubenswrapper[4837]: I1001 09:45:04.789228 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn"] Oct 01 09:45:04 crc kubenswrapper[4837]: I1001 09:45:04.802490 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321820-rg2tn"] Oct 01 09:45:05 crc kubenswrapper[4837]: I1001 09:45:05.842035 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d2727e4-0222-4d79-b696-eff0350a53eb" path="/var/lib/kubelet/pods/2d2727e4-0222-4d79-b696-eff0350a53eb/volumes" Oct 01 09:45:16 crc kubenswrapper[4837]: I1001 09:45:16.168064 4837 scope.go:117] "RemoveContainer" containerID="2136ad3faee1291bcf38372fc195fe002b174a7830e0534ec0f67dfa651a1954" Oct 01 09:45:23 crc kubenswrapper[4837]: I1001 09:45:23.079108 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:45:23 crc kubenswrapper[4837]: I1001 09:45:23.081078 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.079224 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.079910 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.079956 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.080928 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.081023 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7" gracePeriod=600 Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.905568 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7" exitCode=0 Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.905618 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7"} Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.906419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394"} Oct 01 09:45:53 crc kubenswrapper[4837]: I1001 09:45:53.906454 4837 scope.go:117] "RemoveContainer" containerID="b56c2ff8d7e514deb3b939906e8449ca6c2b76c09e1fe33be017d96e3ba81172" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.232998 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:09 crc kubenswrapper[4837]: E1001 09:46:09.233980 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ead52-9298-441f-bca2-7fadf774448b" containerName="collect-profiles" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.233995 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ead52-9298-441f-bca2-7fadf774448b" containerName="collect-profiles" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.234216 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ead52-9298-441f-bca2-7fadf774448b" containerName="collect-profiles" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.236589 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.249541 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.387213 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.388198 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrv2r\" (UniqueName: \"kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.388336 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.492605 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.492741 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrv2r\" (UniqueName: \"kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.492866 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.493756 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.494077 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.531973 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrv2r\" (UniqueName: \"kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r\") pod \"community-operators-nvl88\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:09 crc kubenswrapper[4837]: I1001 09:46:09.565222 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:10 crc kubenswrapper[4837]: I1001 09:46:10.140055 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:10 crc kubenswrapper[4837]: W1001 09:46:10.141215 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51fe7399_69e2_4b33_9760_190591f05b55.slice/crio-76a91165f1c98c458f0eb9673ad4c370beb65bf5e5e8ba0c05ddd49e00b52c80 WatchSource:0}: Error finding container 76a91165f1c98c458f0eb9673ad4c370beb65bf5e5e8ba0c05ddd49e00b52c80: Status 404 returned error can't find the container with id 76a91165f1c98c458f0eb9673ad4c370beb65bf5e5e8ba0c05ddd49e00b52c80 Oct 01 09:46:11 crc kubenswrapper[4837]: I1001 09:46:11.132608 4837 generic.go:334] "Generic (PLEG): container finished" podID="51fe7399-69e2-4b33-9760-190591f05b55" containerID="c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca" exitCode=0 Oct 01 09:46:11 crc kubenswrapper[4837]: I1001 09:46:11.132675 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerDied","Data":"c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca"} Oct 01 09:46:11 crc kubenswrapper[4837]: I1001 09:46:11.132868 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerStarted","Data":"76a91165f1c98c458f0eb9673ad4c370beb65bf5e5e8ba0c05ddd49e00b52c80"} Oct 01 09:46:13 crc kubenswrapper[4837]: I1001 09:46:13.157085 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerStarted","Data":"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4"} Oct 01 09:46:15 crc kubenswrapper[4837]: I1001 09:46:15.182375 4837 generic.go:334] "Generic (PLEG): container finished" podID="51fe7399-69e2-4b33-9760-190591f05b55" containerID="f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4" exitCode=0 Oct 01 09:46:15 crc kubenswrapper[4837]: I1001 09:46:15.182446 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerDied","Data":"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4"} Oct 01 09:46:16 crc kubenswrapper[4837]: I1001 09:46:16.196253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerStarted","Data":"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13"} Oct 01 09:46:16 crc kubenswrapper[4837]: I1001 09:46:16.237596 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nvl88" podStartSLOduration=2.780892368 podStartE2EDuration="7.237577592s" podCreationTimestamp="2025-10-01 09:46:09 +0000 UTC" firstStartedPulling="2025-10-01 09:46:11.135081994 +0000 UTC m=+9627.976689459" lastFinishedPulling="2025-10-01 09:46:15.591767228 +0000 UTC m=+9632.433374683" observedRunningTime="2025-10-01 09:46:16.223301829 +0000 UTC m=+9633.064909284" watchObservedRunningTime="2025-10-01 09:46:16.237577592 +0000 UTC m=+9633.079185037" Oct 01 09:46:19 crc kubenswrapper[4837]: I1001 09:46:19.565890 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:19 crc kubenswrapper[4837]: I1001 09:46:19.566288 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:20 crc kubenswrapper[4837]: I1001 09:46:20.624170 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nvl88" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="registry-server" probeResult="failure" output=< Oct 01 09:46:20 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:46:20 crc kubenswrapper[4837]: > Oct 01 09:46:29 crc kubenswrapper[4837]: I1001 09:46:29.641501 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:29 crc kubenswrapper[4837]: I1001 09:46:29.704056 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:29 crc kubenswrapper[4837]: I1001 09:46:29.885477 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:31 crc kubenswrapper[4837]: I1001 09:46:31.388298 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nvl88" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="registry-server" containerID="cri-o://a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13" gracePeriod=2 Oct 01 09:46:31 crc kubenswrapper[4837]: I1001 09:46:31.882039 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.024357 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrv2r\" (UniqueName: \"kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r\") pod \"51fe7399-69e2-4b33-9760-190591f05b55\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.024444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities\") pod \"51fe7399-69e2-4b33-9760-190591f05b55\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.024777 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content\") pod \"51fe7399-69e2-4b33-9760-190591f05b55\" (UID: \"51fe7399-69e2-4b33-9760-190591f05b55\") " Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.025268 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities" (OuterVolumeSpecName: "utilities") pod "51fe7399-69e2-4b33-9760-190591f05b55" (UID: "51fe7399-69e2-4b33-9760-190591f05b55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.025592 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.032113 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r" (OuterVolumeSpecName: "kube-api-access-xrv2r") pod "51fe7399-69e2-4b33-9760-190591f05b55" (UID: "51fe7399-69e2-4b33-9760-190591f05b55"). InnerVolumeSpecName "kube-api-access-xrv2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.082865 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51fe7399-69e2-4b33-9760-190591f05b55" (UID: "51fe7399-69e2-4b33-9760-190591f05b55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.127725 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fe7399-69e2-4b33-9760-190591f05b55-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.127995 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrv2r\" (UniqueName: \"kubernetes.io/projected/51fe7399-69e2-4b33-9760-190591f05b55-kube-api-access-xrv2r\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.409954 4837 generic.go:334] "Generic (PLEG): container finished" podID="51fe7399-69e2-4b33-9760-190591f05b55" containerID="a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13" exitCode=0 Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.410006 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerDied","Data":"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13"} Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.410036 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvl88" event={"ID":"51fe7399-69e2-4b33-9760-190591f05b55","Type":"ContainerDied","Data":"76a91165f1c98c458f0eb9673ad4c370beb65bf5e5e8ba0c05ddd49e00b52c80"} Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.410056 4837 scope.go:117] "RemoveContainer" containerID="a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.410227 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvl88" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.453641 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.458039 4837 scope.go:117] "RemoveContainer" containerID="f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.466246 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nvl88"] Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.494502 4837 scope.go:117] "RemoveContainer" containerID="c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.541996 4837 scope.go:117] "RemoveContainer" containerID="a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13" Oct 01 09:46:32 crc kubenswrapper[4837]: E1001 09:46:32.542835 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13\": container with ID starting with a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13 not found: ID does not exist" containerID="a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.542884 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13"} err="failed to get container status \"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13\": rpc error: code = NotFound desc = could not find container \"a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13\": container with ID starting with a3f0cacca1df737ba1eb3146b38871771972a79ceedf9c83fedb75a7939e9f13 not found: ID does not exist" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.542911 4837 scope.go:117] "RemoveContainer" containerID="f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4" Oct 01 09:46:32 crc kubenswrapper[4837]: E1001 09:46:32.543371 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4\": container with ID starting with f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4 not found: ID does not exist" containerID="f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.543406 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4"} err="failed to get container status \"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4\": rpc error: code = NotFound desc = could not find container \"f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4\": container with ID starting with f235a9d029731bee752dcc21acbd4d3137a13980997bd5885341993e1994a1f4 not found: ID does not exist" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.543425 4837 scope.go:117] "RemoveContainer" containerID="c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca" Oct 01 09:46:32 crc kubenswrapper[4837]: E1001 09:46:32.544159 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca\": container with ID starting with c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca not found: ID does not exist" containerID="c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca" Oct 01 09:46:32 crc kubenswrapper[4837]: I1001 09:46:32.544219 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca"} err="failed to get container status \"c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca\": rpc error: code = NotFound desc = could not find container \"c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca\": container with ID starting with c4df4f482c8dc4144548adff95087ab9be95fd5112b46d03dbce3edc6c2e8dca not found: ID does not exist" Oct 01 09:46:33 crc kubenswrapper[4837]: I1001 09:46:33.833917 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fe7399-69e2-4b33-9760-190591f05b55" path="/var/lib/kubelet/pods/51fe7399-69e2-4b33-9760-190591f05b55/volumes" Oct 01 09:47:48 crc kubenswrapper[4837]: I1001 09:47:48.350853 4837 generic.go:334] "Generic (PLEG): container finished" podID="58f455af-4ff9-4beb-a59f-8ccfc4860c67" containerID="43c6aec58b562edfa55d55bce68960a9f3a2abc48ae525a70e6d4ab4adc0f428" exitCode=0 Oct 01 09:47:48 crc kubenswrapper[4837]: I1001 09:47:48.351000 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" event={"ID":"58f455af-4ff9-4beb-a59f-8ccfc4860c67","Type":"ContainerDied","Data":"43c6aec58b562edfa55d55bce68960a9f3a2abc48ae525a70e6d4ab4adc0f428"} Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.853480 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926435 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926557 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926579 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926726 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926758 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926776 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926803 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bcqp\" (UniqueName: \"kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.926819 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1\") pod \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\" (UID: \"58f455af-4ff9-4beb-a59f-8ccfc4860c67\") " Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.941171 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp" (OuterVolumeSpecName: "kube-api-access-2bcqp") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "kube-api-access-2bcqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.942144 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.963617 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.971432 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.975491 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.985736 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.986050 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.987544 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory" (OuterVolumeSpecName: "inventory") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:49 crc kubenswrapper[4837]: I1001 09:47:49.993031 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "58f455af-4ff9-4beb-a59f-8ccfc4860c67" (UID: "58f455af-4ff9-4beb-a59f-8ccfc4860c67"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028825 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028860 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028871 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028881 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028890 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028901 4837 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028910 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028920 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bcqp\" (UniqueName: \"kubernetes.io/projected/58f455af-4ff9-4beb-a59f-8ccfc4860c67-kube-api-access-2bcqp\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.028928 4837 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/58f455af-4ff9-4beb-a59f-8ccfc4860c67-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.374276 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" event={"ID":"58f455af-4ff9-4beb-a59f-8ccfc4860c67","Type":"ContainerDied","Data":"308eac78948aed843d877e7f6383d062aeecc5ac8b540212ad49a34fdcfd461a"} Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.374328 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="308eac78948aed843d877e7f6383d062aeecc5ac8b540212ad49a34fdcfd461a" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.374356 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-rjp2f" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.511399 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-dg7g9"] Oct 01 09:47:50 crc kubenswrapper[4837]: E1001 09:47:50.511864 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="extract-content" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.511882 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="extract-content" Oct 01 09:47:50 crc kubenswrapper[4837]: E1001 09:47:50.511909 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="registry-server" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.511917 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="registry-server" Oct 01 09:47:50 crc kubenswrapper[4837]: E1001 09:47:50.511952 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="extract-utilities" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.511962 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="extract-utilities" Oct 01 09:47:50 crc kubenswrapper[4837]: E1001 09:47:50.511980 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f455af-4ff9-4beb-a59f-8ccfc4860c67" containerName="nova-cell1-openstack-openstack-cell1" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.511991 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f455af-4ff9-4beb-a59f-8ccfc4860c67" containerName="nova-cell1-openstack-openstack-cell1" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.512226 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f455af-4ff9-4beb-a59f-8ccfc4860c67" containerName="nova-cell1-openstack-openstack-cell1" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.512250 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fe7399-69e2-4b33-9760-190591f05b55" containerName="registry-server" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.513143 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.515651 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.515907 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.516079 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.516242 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.518079 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.536875 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-dg7g9"] Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.642685 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643099 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643242 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643561 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbrts\" (UniqueName: \"kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643771 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.643903 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.745573 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.745735 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbrts\" (UniqueName: \"kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.745826 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.745855 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.745925 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.746057 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.746104 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.752798 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.753155 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.753881 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.754019 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.755841 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.756139 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.773949 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbrts\" (UniqueName: \"kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts\") pod \"telemetry-openstack-openstack-cell1-dg7g9\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:50 crc kubenswrapper[4837]: I1001 09:47:50.832298 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:47:51 crc kubenswrapper[4837]: I1001 09:47:51.463942 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-dg7g9"] Oct 01 09:47:51 crc kubenswrapper[4837]: W1001 09:47:51.477063 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod177a9873_846f_4c66_8dda_abe9d4874b96.slice/crio-c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9 WatchSource:0}: Error finding container c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9: Status 404 returned error can't find the container with id c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9 Oct 01 09:47:52 crc kubenswrapper[4837]: I1001 09:47:52.399664 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" event={"ID":"177a9873-846f-4c66-8dda-abe9d4874b96","Type":"ContainerStarted","Data":"87a3869bea795a4f9912c2e41421f8e6fa98b175afcc0fdf697638582c5ed8a8"} Oct 01 09:47:52 crc kubenswrapper[4837]: I1001 09:47:52.400073 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" event={"ID":"177a9873-846f-4c66-8dda-abe9d4874b96","Type":"ContainerStarted","Data":"c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9"} Oct 01 09:47:52 crc kubenswrapper[4837]: I1001 09:47:52.425999 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" podStartSLOduration=1.980476227 podStartE2EDuration="2.425968444s" podCreationTimestamp="2025-10-01 09:47:50 +0000 UTC" firstStartedPulling="2025-10-01 09:47:51.479398106 +0000 UTC m=+9728.321005571" lastFinishedPulling="2025-10-01 09:47:51.924890283 +0000 UTC m=+9728.766497788" observedRunningTime="2025-10-01 09:47:52.416418828 +0000 UTC m=+9729.258026293" watchObservedRunningTime="2025-10-01 09:47:52.425968444 +0000 UTC m=+9729.267575939" Oct 01 09:47:53 crc kubenswrapper[4837]: I1001 09:47:53.079093 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:47:53 crc kubenswrapper[4837]: I1001 09:47:53.079550 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:48:23 crc kubenswrapper[4837]: I1001 09:48:23.079167 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:48:23 crc kubenswrapper[4837]: I1001 09:48:23.079709 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:48:53 crc kubenswrapper[4837]: I1001 09:48:53.080234 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:48:53 crc kubenswrapper[4837]: I1001 09:48:53.080959 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:48:53 crc kubenswrapper[4837]: I1001 09:48:53.081022 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:48:53 crc kubenswrapper[4837]: I1001 09:48:53.082303 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:48:53 crc kubenswrapper[4837]: I1001 09:48:53.082409 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" gracePeriod=600 Oct 01 09:48:53 crc kubenswrapper[4837]: E1001 09:48:53.203123 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:48:54 crc kubenswrapper[4837]: I1001 09:48:54.079270 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" exitCode=0 Oct 01 09:48:54 crc kubenswrapper[4837]: I1001 09:48:54.079364 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394"} Oct 01 09:48:54 crc kubenswrapper[4837]: I1001 09:48:54.079665 4837 scope.go:117] "RemoveContainer" containerID="c5d15433c2e1a136d61b470a9fa3dc2cc69234211770baf12c9126f91530cfc7" Oct 01 09:48:54 crc kubenswrapper[4837]: I1001 09:48:54.080466 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:48:54 crc kubenswrapper[4837]: E1001 09:48:54.080846 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:49:08 crc kubenswrapper[4837]: I1001 09:49:08.816152 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:49:08 crc kubenswrapper[4837]: E1001 09:49:08.817027 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:49:21 crc kubenswrapper[4837]: I1001 09:49:21.816949 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:49:21 crc kubenswrapper[4837]: E1001 09:49:21.818167 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:49:33 crc kubenswrapper[4837]: I1001 09:49:33.698250 4837 trace.go:236] Trace[340696088]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-j6jvg" (01-Oct-2025 09:49:32.689) (total time: 1008ms): Oct 01 09:49:33 crc kubenswrapper[4837]: Trace[340696088]: [1.008705363s] [1.008705363s] END Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.640346 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.642545 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.658201 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.751322 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.751381 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.751826 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9hlh\" (UniqueName: \"kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.815725 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:49:34 crc kubenswrapper[4837]: E1001 09:49:34.815987 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.853788 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.853852 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.853949 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9hlh\" (UniqueName: \"kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.854458 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.854524 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.876720 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9hlh\" (UniqueName: \"kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh\") pod \"redhat-marketplace-5rxr7\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:34 crc kubenswrapper[4837]: I1001 09:49:34.967269 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:35 crc kubenswrapper[4837]: I1001 09:49:35.441201 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:36 crc kubenswrapper[4837]: I1001 09:49:36.565633 4837 generic.go:334] "Generic (PLEG): container finished" podID="06687683-6198-4df3-a468-87012e1d5d1d" containerID="45ba1980824781e597f2b74403915d7e2c9b7dc48ac034277612e88a695fce12" exitCode=0 Oct 01 09:49:36 crc kubenswrapper[4837]: I1001 09:49:36.565844 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerDied","Data":"45ba1980824781e597f2b74403915d7e2c9b7dc48ac034277612e88a695fce12"} Oct 01 09:49:36 crc kubenswrapper[4837]: I1001 09:49:36.566442 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerStarted","Data":"60fc0c968f3ecc6ec7f2a14768fb991ab9769b6851658d2e08709922be7c3c37"} Oct 01 09:49:36 crc kubenswrapper[4837]: I1001 09:49:36.569193 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:49:37 crc kubenswrapper[4837]: I1001 09:49:37.577498 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerStarted","Data":"0e0ac6c4d59247bba01c302a0804a667744d4f6baa7acef55973e738a5237c89"} Oct 01 09:49:38 crc kubenswrapper[4837]: I1001 09:49:38.593537 4837 generic.go:334] "Generic (PLEG): container finished" podID="06687683-6198-4df3-a468-87012e1d5d1d" containerID="0e0ac6c4d59247bba01c302a0804a667744d4f6baa7acef55973e738a5237c89" exitCode=0 Oct 01 09:49:38 crc kubenswrapper[4837]: I1001 09:49:38.593649 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerDied","Data":"0e0ac6c4d59247bba01c302a0804a667744d4f6baa7acef55973e738a5237c89"} Oct 01 09:49:39 crc kubenswrapper[4837]: I1001 09:49:39.608657 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerStarted","Data":"debf440c5f758d2981004b728594f4bc822a928a7e2e976ce242c48ff674b39e"} Oct 01 09:49:39 crc kubenswrapper[4837]: I1001 09:49:39.629095 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5rxr7" podStartSLOduration=3.02138075 podStartE2EDuration="5.629073667s" podCreationTimestamp="2025-10-01 09:49:34 +0000 UTC" firstStartedPulling="2025-10-01 09:49:36.568516319 +0000 UTC m=+9833.410123814" lastFinishedPulling="2025-10-01 09:49:39.176209256 +0000 UTC m=+9836.017816731" observedRunningTime="2025-10-01 09:49:39.626131594 +0000 UTC m=+9836.467739049" watchObservedRunningTime="2025-10-01 09:49:39.629073667 +0000 UTC m=+9836.470681122" Oct 01 09:49:44 crc kubenswrapper[4837]: I1001 09:49:44.967861 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:44 crc kubenswrapper[4837]: I1001 09:49:44.968591 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:45 crc kubenswrapper[4837]: I1001 09:49:45.028141 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:46 crc kubenswrapper[4837]: I1001 09:49:46.346262 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:46 crc kubenswrapper[4837]: I1001 09:49:46.396043 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:46 crc kubenswrapper[4837]: I1001 09:49:46.816732 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:49:46 crc kubenswrapper[4837]: E1001 09:49:46.817073 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:49:47 crc kubenswrapper[4837]: I1001 09:49:47.692805 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5rxr7" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="registry-server" containerID="cri-o://debf440c5f758d2981004b728594f4bc822a928a7e2e976ce242c48ff674b39e" gracePeriod=2 Oct 01 09:49:48 crc kubenswrapper[4837]: I1001 09:49:48.702915 4837 generic.go:334] "Generic (PLEG): container finished" podID="06687683-6198-4df3-a468-87012e1d5d1d" containerID="debf440c5f758d2981004b728594f4bc822a928a7e2e976ce242c48ff674b39e" exitCode=0 Oct 01 09:49:48 crc kubenswrapper[4837]: I1001 09:49:48.703026 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerDied","Data":"debf440c5f758d2981004b728594f4bc822a928a7e2e976ce242c48ff674b39e"} Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.431942 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.490087 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities\") pod \"06687683-6198-4df3-a468-87012e1d5d1d\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.490203 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content\") pod \"06687683-6198-4df3-a468-87012e1d5d1d\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.490317 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9hlh\" (UniqueName: \"kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh\") pod \"06687683-6198-4df3-a468-87012e1d5d1d\" (UID: \"06687683-6198-4df3-a468-87012e1d5d1d\") " Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.492155 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities" (OuterVolumeSpecName: "utilities") pod "06687683-6198-4df3-a468-87012e1d5d1d" (UID: "06687683-6198-4df3-a468-87012e1d5d1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.497635 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh" (OuterVolumeSpecName: "kube-api-access-h9hlh") pod "06687683-6198-4df3-a468-87012e1d5d1d" (UID: "06687683-6198-4df3-a468-87012e1d5d1d"). InnerVolumeSpecName "kube-api-access-h9hlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.503159 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06687683-6198-4df3-a468-87012e1d5d1d" (UID: "06687683-6198-4df3-a468-87012e1d5d1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.593060 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.593114 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06687683-6198-4df3-a468-87012e1d5d1d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.593131 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9hlh\" (UniqueName: \"kubernetes.io/projected/06687683-6198-4df3-a468-87012e1d5d1d-kube-api-access-h9hlh\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.718924 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rxr7" event={"ID":"06687683-6198-4df3-a468-87012e1d5d1d","Type":"ContainerDied","Data":"60fc0c968f3ecc6ec7f2a14768fb991ab9769b6851658d2e08709922be7c3c37"} Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.719267 4837 scope.go:117] "RemoveContainer" containerID="debf440c5f758d2981004b728594f4bc822a928a7e2e976ce242c48ff674b39e" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.719002 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rxr7" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.765074 4837 scope.go:117] "RemoveContainer" containerID="0e0ac6c4d59247bba01c302a0804a667744d4f6baa7acef55973e738a5237c89" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.767461 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.781763 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rxr7"] Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.796053 4837 scope.go:117] "RemoveContainer" containerID="45ba1980824781e597f2b74403915d7e2c9b7dc48ac034277612e88a695fce12" Oct 01 09:49:49 crc kubenswrapper[4837]: I1001 09:49:49.831205 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06687683-6198-4df3-a468-87012e1d5d1d" path="/var/lib/kubelet/pods/06687683-6198-4df3-a468-87012e1d5d1d/volumes" Oct 01 09:50:01 crc kubenswrapper[4837]: I1001 09:50:01.817064 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:50:01 crc kubenswrapper[4837]: E1001 09:50:01.817944 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:50:16 crc kubenswrapper[4837]: I1001 09:50:16.816766 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:50:16 crc kubenswrapper[4837]: E1001 09:50:16.817902 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:50:28 crc kubenswrapper[4837]: I1001 09:50:28.816055 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:50:28 crc kubenswrapper[4837]: E1001 09:50:28.817430 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:50:43 crc kubenswrapper[4837]: I1001 09:50:43.836004 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:50:43 crc kubenswrapper[4837]: E1001 09:50:43.837421 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:50:58 crc kubenswrapper[4837]: I1001 09:50:58.816538 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:50:58 crc kubenswrapper[4837]: E1001 09:50:58.817833 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:51:12 crc kubenswrapper[4837]: I1001 09:51:12.817383 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:51:12 crc kubenswrapper[4837]: E1001 09:51:12.818540 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.984896 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:21 crc kubenswrapper[4837]: E1001 09:51:21.985993 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="extract-utilities" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.986090 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="extract-utilities" Oct 01 09:51:21 crc kubenswrapper[4837]: E1001 09:51:21.986109 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="extract-content" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.986118 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="extract-content" Oct 01 09:51:21 crc kubenswrapper[4837]: E1001 09:51:21.986155 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="registry-server" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.986164 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="registry-server" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.986454 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="06687683-6198-4df3-a468-87012e1d5d1d" containerName="registry-server" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.988338 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:21 crc kubenswrapper[4837]: I1001 09:51:21.993576 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.058459 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.058817 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2m6\" (UniqueName: \"kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.059074 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.160607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.160669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.160848 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2m6\" (UniqueName: \"kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.161661 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.162031 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.192222 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2m6\" (UniqueName: \"kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6\") pod \"redhat-operators-tggzr\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.323927 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:22 crc kubenswrapper[4837]: I1001 09:51:22.877406 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:23 crc kubenswrapper[4837]: I1001 09:51:23.861765 4837 generic.go:334] "Generic (PLEG): container finished" podID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerID="a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433" exitCode=0 Oct 01 09:51:23 crc kubenswrapper[4837]: I1001 09:51:23.862031 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerDied","Data":"a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433"} Oct 01 09:51:23 crc kubenswrapper[4837]: I1001 09:51:23.862058 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerStarted","Data":"a78aee7daf2db8c54a783063c180f522b88fdfeaa31910bc730fb630922e062f"} Oct 01 09:51:24 crc kubenswrapper[4837]: I1001 09:51:24.877192 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerStarted","Data":"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28"} Oct 01 09:51:25 crc kubenswrapper[4837]: I1001 09:51:25.817036 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:51:25 crc kubenswrapper[4837]: E1001 09:51:25.817489 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:51:29 crc kubenswrapper[4837]: I1001 09:51:29.945828 4837 generic.go:334] "Generic (PLEG): container finished" podID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerID="cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28" exitCode=0 Oct 01 09:51:29 crc kubenswrapper[4837]: I1001 09:51:29.946633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerDied","Data":"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28"} Oct 01 09:51:30 crc kubenswrapper[4837]: I1001 09:51:30.959677 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerStarted","Data":"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f"} Oct 01 09:51:30 crc kubenswrapper[4837]: I1001 09:51:30.984169 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tggzr" podStartSLOduration=3.4749589 podStartE2EDuration="9.984147884s" podCreationTimestamp="2025-10-01 09:51:21 +0000 UTC" firstStartedPulling="2025-10-01 09:51:23.864922389 +0000 UTC m=+9940.706529854" lastFinishedPulling="2025-10-01 09:51:30.374111383 +0000 UTC m=+9947.215718838" observedRunningTime="2025-10-01 09:51:30.978429773 +0000 UTC m=+9947.820037238" watchObservedRunningTime="2025-10-01 09:51:30.984147884 +0000 UTC m=+9947.825755349" Oct 01 09:51:32 crc kubenswrapper[4837]: I1001 09:51:32.325022 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:32 crc kubenswrapper[4837]: I1001 09:51:32.325574 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:33 crc kubenswrapper[4837]: I1001 09:51:33.389821 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tggzr" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" probeResult="failure" output=< Oct 01 09:51:33 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:51:33 crc kubenswrapper[4837]: > Oct 01 09:51:40 crc kubenswrapper[4837]: I1001 09:51:40.832024 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:51:40 crc kubenswrapper[4837]: E1001 09:51:40.832839 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:51:43 crc kubenswrapper[4837]: I1001 09:51:43.385471 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tggzr" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" probeResult="failure" output=< Oct 01 09:51:43 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 09:51:43 crc kubenswrapper[4837]: > Oct 01 09:51:52 crc kubenswrapper[4837]: I1001 09:51:52.411178 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:52 crc kubenswrapper[4837]: I1001 09:51:52.491777 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:52 crc kubenswrapper[4837]: I1001 09:51:52.816550 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:51:52 crc kubenswrapper[4837]: E1001 09:51:52.816985 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:51:53 crc kubenswrapper[4837]: I1001 09:51:53.177040 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.252741 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tggzr" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" containerID="cri-o://0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f" gracePeriod=2 Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.772084 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.899567 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities\") pod \"836f58ad-ee67-4107-8be4-cf20d0e1e404\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.899759 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx2m6\" (UniqueName: \"kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6\") pod \"836f58ad-ee67-4107-8be4-cf20d0e1e404\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.899884 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content\") pod \"836f58ad-ee67-4107-8be4-cf20d0e1e404\" (UID: \"836f58ad-ee67-4107-8be4-cf20d0e1e404\") " Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.900802 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities" (OuterVolumeSpecName: "utilities") pod "836f58ad-ee67-4107-8be4-cf20d0e1e404" (UID: "836f58ad-ee67-4107-8be4-cf20d0e1e404"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.901018 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.909251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6" (OuterVolumeSpecName: "kube-api-access-bx2m6") pod "836f58ad-ee67-4107-8be4-cf20d0e1e404" (UID: "836f58ad-ee67-4107-8be4-cf20d0e1e404"). InnerVolumeSpecName "kube-api-access-bx2m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:54 crc kubenswrapper[4837]: I1001 09:51:54.995545 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "836f58ad-ee67-4107-8be4-cf20d0e1e404" (UID: "836f58ad-ee67-4107-8be4-cf20d0e1e404"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.002373 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx2m6\" (UniqueName: \"kubernetes.io/projected/836f58ad-ee67-4107-8be4-cf20d0e1e404-kube-api-access-bx2m6\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.002400 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/836f58ad-ee67-4107-8be4-cf20d0e1e404-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.265265 4837 generic.go:334] "Generic (PLEG): container finished" podID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerID="0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f" exitCode=0 Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.265298 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerDied","Data":"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f"} Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.265603 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tggzr" event={"ID":"836f58ad-ee67-4107-8be4-cf20d0e1e404","Type":"ContainerDied","Data":"a78aee7daf2db8c54a783063c180f522b88fdfeaa31910bc730fb630922e062f"} Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.265627 4837 scope.go:117] "RemoveContainer" containerID="0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.265338 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tggzr" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.295109 4837 scope.go:117] "RemoveContainer" containerID="cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.324999 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.335729 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tggzr"] Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.354315 4837 scope.go:117] "RemoveContainer" containerID="a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.393254 4837 scope.go:117] "RemoveContainer" containerID="0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f" Oct 01 09:51:55 crc kubenswrapper[4837]: E1001 09:51:55.393655 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f\": container with ID starting with 0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f not found: ID does not exist" containerID="0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.393685 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f"} err="failed to get container status \"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f\": rpc error: code = NotFound desc = could not find container \"0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f\": container with ID starting with 0c52aa6c70fbc88b7d7111be08f5866572f139a2e05a8364619ad2a3df6cf70f not found: ID does not exist" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.393932 4837 scope.go:117] "RemoveContainer" containerID="cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28" Oct 01 09:51:55 crc kubenswrapper[4837]: E1001 09:51:55.399616 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28\": container with ID starting with cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28 not found: ID does not exist" containerID="cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.399673 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28"} err="failed to get container status \"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28\": rpc error: code = NotFound desc = could not find container \"cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28\": container with ID starting with cd949316e5f43546da8ebd3737eff95771254bed7c355b88b2749f2784090b28 not found: ID does not exist" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.399724 4837 scope.go:117] "RemoveContainer" containerID="a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433" Oct 01 09:51:55 crc kubenswrapper[4837]: E1001 09:51:55.400308 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433\": container with ID starting with a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433 not found: ID does not exist" containerID="a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.400362 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433"} err="failed to get container status \"a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433\": rpc error: code = NotFound desc = could not find container \"a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433\": container with ID starting with a58e10cabdbec78caf93b30824ef4de1e01ee64101b510f89556946227e5e433 not found: ID does not exist" Oct 01 09:51:55 crc kubenswrapper[4837]: E1001 09:51:55.438803 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod836f58ad_ee67_4107_8be4_cf20d0e1e404.slice/crio-a78aee7daf2db8c54a783063c180f522b88fdfeaa31910bc730fb630922e062f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod836f58ad_ee67_4107_8be4_cf20d0e1e404.slice\": RecentStats: unable to find data in memory cache]" Oct 01 09:51:55 crc kubenswrapper[4837]: I1001 09:51:55.832542 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" path="/var/lib/kubelet/pods/836f58ad-ee67-4107-8be4-cf20d0e1e404/volumes" Oct 01 09:52:04 crc kubenswrapper[4837]: I1001 09:52:04.817090 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:52:04 crc kubenswrapper[4837]: E1001 09:52:04.818208 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:52:16 crc kubenswrapper[4837]: I1001 09:52:16.816521 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:52:16 crc kubenswrapper[4837]: E1001 09:52:16.817497 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:52:27 crc kubenswrapper[4837]: I1001 09:52:27.815979 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:52:27 crc kubenswrapper[4837]: E1001 09:52:27.816776 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:52:40 crc kubenswrapper[4837]: I1001 09:52:40.816451 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:52:40 crc kubenswrapper[4837]: E1001 09:52:40.817243 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:52:55 crc kubenswrapper[4837]: I1001 09:52:55.816682 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:52:55 crc kubenswrapper[4837]: E1001 09:52:55.817978 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:53:06 crc kubenswrapper[4837]: I1001 09:53:06.816398 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:53:06 crc kubenswrapper[4837]: E1001 09:53:06.817524 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:53:18 crc kubenswrapper[4837]: I1001 09:53:18.816301 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:53:18 crc kubenswrapper[4837]: E1001 09:53:18.817363 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:53:31 crc kubenswrapper[4837]: I1001 09:53:31.816296 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:53:31 crc kubenswrapper[4837]: E1001 09:53:31.817428 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:53:43 crc kubenswrapper[4837]: I1001 09:53:43.835091 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:53:43 crc kubenswrapper[4837]: E1001 09:53:43.836425 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 09:53:56 crc kubenswrapper[4837]: I1001 09:53:56.816945 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:53:57 crc kubenswrapper[4837]: I1001 09:53:57.779942 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157"} Oct 01 09:55:17 crc kubenswrapper[4837]: I1001 09:55:17.809523 4837 generic.go:334] "Generic (PLEG): container finished" podID="177a9873-846f-4c66-8dda-abe9d4874b96" containerID="87a3869bea795a4f9912c2e41421f8e6fa98b175afcc0fdf697638582c5ed8a8" exitCode=0 Oct 01 09:55:17 crc kubenswrapper[4837]: I1001 09:55:17.809576 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" event={"ID":"177a9873-846f-4c66-8dda-abe9d4874b96","Type":"ContainerDied","Data":"87a3869bea795a4f9912c2e41421f8e6fa98b175afcc0fdf697638582c5ed8a8"} Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.397714 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539247 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbrts\" (UniqueName: \"kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539324 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539410 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539474 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539581 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539747 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.539795 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key\") pod \"177a9873-846f-4c66-8dda-abe9d4874b96\" (UID: \"177a9873-846f-4c66-8dda-abe9d4874b96\") " Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.547003 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts" (OuterVolumeSpecName: "kube-api-access-fbrts") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "kube-api-access-fbrts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.560525 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.578610 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.586501 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.588951 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory" (OuterVolumeSpecName: "inventory") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.589585 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.605247 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "177a9873-846f-4c66-8dda-abe9d4874b96" (UID: "177a9873-846f-4c66-8dda-abe9d4874b96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.642952 4837 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643000 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643016 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbrts\" (UniqueName: \"kubernetes.io/projected/177a9873-846f-4c66-8dda-abe9d4874b96-kube-api-access-fbrts\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643027 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643040 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643061 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.643073 4837 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/177a9873-846f-4c66-8dda-abe9d4874b96-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.842049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" event={"ID":"177a9873-846f-4c66-8dda-abe9d4874b96","Type":"ContainerDied","Data":"c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9"} Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.842498 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1df0c03f81154f2f46a360264882c067f2b7f30037109627ec77e30f7dd28c9" Oct 01 09:55:19 crc kubenswrapper[4837]: I1001 09:55:19.842114 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-dg7g9" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.143251 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-zgx57"] Oct 01 09:55:20 crc kubenswrapper[4837]: E1001 09:55:20.144349 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="extract-utilities" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.144528 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="extract-utilities" Oct 01 09:55:20 crc kubenswrapper[4837]: E1001 09:55:20.144683 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="extract-content" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.144845 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="extract-content" Oct 01 09:55:20 crc kubenswrapper[4837]: E1001 09:55:20.145015 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.145146 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" Oct 01 09:55:20 crc kubenswrapper[4837]: E1001 09:55:20.145304 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177a9873-846f-4c66-8dda-abe9d4874b96" containerName="telemetry-openstack-openstack-cell1" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.145433 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="177a9873-846f-4c66-8dda-abe9d4874b96" containerName="telemetry-openstack-openstack-cell1" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.145939 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="836f58ad-ee67-4107-8be4-cf20d0e1e404" containerName="registry-server" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.146113 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="177a9873-846f-4c66-8dda-abe9d4874b96" containerName="telemetry-openstack-openstack-cell1" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.147651 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.149818 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.152037 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.152357 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.152882 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.158125 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-zgx57"] Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.160769 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.261432 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.261502 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.261665 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcrt7\" (UniqueName: \"kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.261762 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.261952 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.363876 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcrt7\" (UniqueName: \"kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.363941 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.364115 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.364228 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.364329 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.370296 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.370329 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.372791 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.375080 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.387176 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcrt7\" (UniqueName: \"kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7\") pod \"neutron-sriov-openstack-openstack-cell1-zgx57\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.496968 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.896030 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-zgx57"] Oct 01 09:55:20 crc kubenswrapper[4837]: I1001 09:55:20.901013 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:55:21 crc kubenswrapper[4837]: I1001 09:55:21.883834 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" event={"ID":"a8855094-af9a-42ba-9931-55420616b36d","Type":"ContainerStarted","Data":"450e2c1b466c4e0509968103ac3f69e2cbe1c419e6f355b96c9f99c2098957e4"} Oct 01 09:55:21 crc kubenswrapper[4837]: I1001 09:55:21.884279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" event={"ID":"a8855094-af9a-42ba-9931-55420616b36d","Type":"ContainerStarted","Data":"05dd8f4672138faaedfd9e51202442db0f62ed477e6dd3036ee488f86481b68b"} Oct 01 09:55:21 crc kubenswrapper[4837]: I1001 09:55:21.918729 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" podStartSLOduration=1.457745553 podStartE2EDuration="1.918708073s" podCreationTimestamp="2025-10-01 09:55:20 +0000 UTC" firstStartedPulling="2025-10-01 09:55:20.900827424 +0000 UTC m=+10177.742434879" lastFinishedPulling="2025-10-01 09:55:21.361789914 +0000 UTC m=+10178.203397399" observedRunningTime="2025-10-01 09:55:21.906468741 +0000 UTC m=+10178.748076236" watchObservedRunningTime="2025-10-01 09:55:21.918708073 +0000 UTC m=+10178.760315528" Oct 01 09:56:07 crc kubenswrapper[4837]: I1001 09:56:07.457812 4837 generic.go:334] "Generic (PLEG): container finished" podID="a8855094-af9a-42ba-9931-55420616b36d" containerID="450e2c1b466c4e0509968103ac3f69e2cbe1c419e6f355b96c9f99c2098957e4" exitCode=0 Oct 01 09:56:07 crc kubenswrapper[4837]: I1001 09:56:07.457909 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" event={"ID":"a8855094-af9a-42ba-9931-55420616b36d","Type":"ContainerDied","Data":"450e2c1b466c4e0509968103ac3f69e2cbe1c419e6f355b96c9f99c2098957e4"} Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.008992 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.065008 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.065109 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.065241 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.065315 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.065411 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcrt7\" (UniqueName: \"kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.073797 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.081240 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7" (OuterVolumeSpecName: "kube-api-access-fcrt7") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d"). InnerVolumeSpecName "kube-api-access-fcrt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:09 crc kubenswrapper[4837]: E1001 09:56:09.099466 4837 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0 podName:a8855094-af9a-42ba-9931-55420616b36d nodeName:}" failed. No retries permitted until 2025-10-01 09:56:09.599435841 +0000 UTC m=+10226.441043286 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "neutron-sriov-agent-neutron-config-0" (UniqueName: "kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d") : error deleting /var/lib/kubelet/pods/a8855094-af9a-42ba-9931-55420616b36d/volume-subpaths: remove /var/lib/kubelet/pods/a8855094-af9a-42ba-9931-55420616b36d/volume-subpaths: no such file or directory Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.101718 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory" (OuterVolumeSpecName: "inventory") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.110789 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.167658 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.167720 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcrt7\" (UniqueName: \"kubernetes.io/projected/a8855094-af9a-42ba-9931-55420616b36d-kube-api-access-fcrt7\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.167737 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.167749 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.482730 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" event={"ID":"a8855094-af9a-42ba-9931-55420616b36d","Type":"ContainerDied","Data":"05dd8f4672138faaedfd9e51202442db0f62ed477e6dd3036ee488f86481b68b"} Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.483184 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05dd8f4672138faaedfd9e51202442db0f62ed477e6dd3036ee488f86481b68b" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.482792 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-zgx57" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.679075 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") pod \"a8855094-af9a-42ba-9931-55420616b36d\" (UID: \"a8855094-af9a-42ba-9931-55420616b36d\") " Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.692044 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "a8855094-af9a-42ba-9931-55420616b36d" (UID: "a8855094-af9a-42ba-9931-55420616b36d"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.701561 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx"] Oct 01 09:56:09 crc kubenswrapper[4837]: E1001 09:56:09.703292 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8855094-af9a-42ba-9931-55420616b36d" containerName="neutron-sriov-openstack-openstack-cell1" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.703313 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8855094-af9a-42ba-9931-55420616b36d" containerName="neutron-sriov-openstack-openstack-cell1" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.703492 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8855094-af9a-42ba-9931-55420616b36d" containerName="neutron-sriov-openstack-openstack-cell1" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.704282 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.706460 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.722779 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx"] Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.782235 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.782378 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzrcc\" (UniqueName: \"kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.782423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.782488 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.782773 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.783079 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a8855094-af9a-42ba-9931-55420616b36d-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.885377 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.885490 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzrcc\" (UniqueName: \"kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.885567 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.885605 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.885654 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.888743 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.889665 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.890124 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.889612 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:09 crc kubenswrapper[4837]: I1001 09:56:09.902869 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzrcc\" (UniqueName: \"kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc\") pod \"neutron-dhcp-openstack-openstack-cell1-2c4mx\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:10 crc kubenswrapper[4837]: I1001 09:56:10.066229 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:56:10 crc kubenswrapper[4837]: I1001 09:56:10.700600 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx"] Oct 01 09:56:11 crc kubenswrapper[4837]: I1001 09:56:11.514901 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" event={"ID":"41ead075-9a6e-4f49-812a-9c8cc8937c22","Type":"ContainerStarted","Data":"d1dabf3d4554758f574b251016374228ecfe1b88334d43c8687d699ef5b22635"} Oct 01 09:56:12 crc kubenswrapper[4837]: I1001 09:56:12.528901 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" event={"ID":"41ead075-9a6e-4f49-812a-9c8cc8937c22","Type":"ContainerStarted","Data":"2ea25d8e8685cb43bdcfbb936e8600e3508cb41c1a0a3559bd4d850d35be85a8"} Oct 01 09:56:12 crc kubenswrapper[4837]: I1001 09:56:12.572210 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" podStartSLOduration=3.115092729 podStartE2EDuration="3.572186924s" podCreationTimestamp="2025-10-01 09:56:09 +0000 UTC" firstStartedPulling="2025-10-01 09:56:10.713160579 +0000 UTC m=+10227.554768064" lastFinishedPulling="2025-10-01 09:56:11.170254764 +0000 UTC m=+10228.011862259" observedRunningTime="2025-10-01 09:56:12.555920012 +0000 UTC m=+10229.397527547" watchObservedRunningTime="2025-10-01 09:56:12.572186924 +0000 UTC m=+10229.413794419" Oct 01 09:56:23 crc kubenswrapper[4837]: I1001 09:56:23.078976 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:56:23 crc kubenswrapper[4837]: I1001 09:56:23.079457 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.854876 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.857434 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.873236 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.973163 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.973259 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:32 crc kubenswrapper[4837]: I1001 09:56:32.973342 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2tsk\" (UniqueName: \"kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.075278 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.075383 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.075465 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2tsk\" (UniqueName: \"kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.075817 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.075872 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.097136 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2tsk\" (UniqueName: \"kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk\") pod \"community-operators-bhqbr\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.181000 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:33 crc kubenswrapper[4837]: I1001 09:56:33.796447 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:34 crc kubenswrapper[4837]: I1001 09:56:34.794270 4837 generic.go:334] "Generic (PLEG): container finished" podID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerID="8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3" exitCode=0 Oct 01 09:56:34 crc kubenswrapper[4837]: I1001 09:56:34.794393 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerDied","Data":"8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3"} Oct 01 09:56:34 crc kubenswrapper[4837]: I1001 09:56:34.794618 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerStarted","Data":"5a015d276c0189fcddd3d5476c8c3f116c9d9853c385fb5f0bad991a038a653f"} Oct 01 09:56:35 crc kubenswrapper[4837]: I1001 09:56:35.808943 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerStarted","Data":"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d"} Oct 01 09:56:37 crc kubenswrapper[4837]: I1001 09:56:37.837907 4837 generic.go:334] "Generic (PLEG): container finished" podID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerID="b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d" exitCode=0 Oct 01 09:56:37 crc kubenswrapper[4837]: I1001 09:56:37.837997 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerDied","Data":"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d"} Oct 01 09:56:39 crc kubenswrapper[4837]: I1001 09:56:39.864443 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerStarted","Data":"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24"} Oct 01 09:56:39 crc kubenswrapper[4837]: I1001 09:56:39.888997 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bhqbr" podStartSLOduration=3.974633858 podStartE2EDuration="7.888978373s" podCreationTimestamp="2025-10-01 09:56:32 +0000 UTC" firstStartedPulling="2025-10-01 09:56:34.799947858 +0000 UTC m=+10251.641555333" lastFinishedPulling="2025-10-01 09:56:38.714292393 +0000 UTC m=+10255.555899848" observedRunningTime="2025-10-01 09:56:39.883448706 +0000 UTC m=+10256.725056171" watchObservedRunningTime="2025-10-01 09:56:39.888978373 +0000 UTC m=+10256.730585838" Oct 01 09:56:43 crc kubenswrapper[4837]: I1001 09:56:43.182763 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:43 crc kubenswrapper[4837]: I1001 09:56:43.183149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:43 crc kubenswrapper[4837]: I1001 09:56:43.273089 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:43 crc kubenswrapper[4837]: I1001 09:56:43.976902 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:44 crc kubenswrapper[4837]: I1001 09:56:44.034077 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:45 crc kubenswrapper[4837]: I1001 09:56:45.935622 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bhqbr" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="registry-server" containerID="cri-o://3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24" gracePeriod=2 Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.569141 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.685105 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2tsk\" (UniqueName: \"kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk\") pod \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.685304 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities\") pod \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.685558 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content\") pod \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\" (UID: \"712274e4-5d3e-4c7b-a779-99fb08d8f5a5\") " Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.686831 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities" (OuterVolumeSpecName: "utilities") pod "712274e4-5d3e-4c7b-a779-99fb08d8f5a5" (UID: "712274e4-5d3e-4c7b-a779-99fb08d8f5a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.696574 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk" (OuterVolumeSpecName: "kube-api-access-r2tsk") pod "712274e4-5d3e-4c7b-a779-99fb08d8f5a5" (UID: "712274e4-5d3e-4c7b-a779-99fb08d8f5a5"). InnerVolumeSpecName "kube-api-access-r2tsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.746191 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "712274e4-5d3e-4c7b-a779-99fb08d8f5a5" (UID: "712274e4-5d3e-4c7b-a779-99fb08d8f5a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.789611 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.789660 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2tsk\" (UniqueName: \"kubernetes.io/projected/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-kube-api-access-r2tsk\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.789682 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/712274e4-5d3e-4c7b-a779-99fb08d8f5a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.955166 4837 generic.go:334] "Generic (PLEG): container finished" podID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerID="3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24" exitCode=0 Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.955288 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bhqbr" Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.955297 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerDied","Data":"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24"} Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.955771 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bhqbr" event={"ID":"712274e4-5d3e-4c7b-a779-99fb08d8f5a5","Type":"ContainerDied","Data":"5a015d276c0189fcddd3d5476c8c3f116c9d9853c385fb5f0bad991a038a653f"} Oct 01 09:56:46 crc kubenswrapper[4837]: I1001 09:56:46.955797 4837 scope.go:117] "RemoveContainer" containerID="3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.001550 4837 scope.go:117] "RemoveContainer" containerID="b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.006366 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.018844 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bhqbr"] Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.047217 4837 scope.go:117] "RemoveContainer" containerID="8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.114795 4837 scope.go:117] "RemoveContainer" containerID="3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24" Oct 01 09:56:47 crc kubenswrapper[4837]: E1001 09:56:47.115679 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24\": container with ID starting with 3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24 not found: ID does not exist" containerID="3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.115739 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24"} err="failed to get container status \"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24\": rpc error: code = NotFound desc = could not find container \"3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24\": container with ID starting with 3ba9f7befa8cbaf255521a314b20e3d01ad861df54ee7253a375c82e3e1d0b24 not found: ID does not exist" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.115777 4837 scope.go:117] "RemoveContainer" containerID="b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d" Oct 01 09:56:47 crc kubenswrapper[4837]: E1001 09:56:47.116196 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d\": container with ID starting with b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d not found: ID does not exist" containerID="b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.116230 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d"} err="failed to get container status \"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d\": rpc error: code = NotFound desc = could not find container \"b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d\": container with ID starting with b8eb1e73b8dc420124eb7407cc3f0d848136e8de2e682c9b1c2e4841cc129e9d not found: ID does not exist" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.116251 4837 scope.go:117] "RemoveContainer" containerID="8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3" Oct 01 09:56:47 crc kubenswrapper[4837]: E1001 09:56:47.116533 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3\": container with ID starting with 8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3 not found: ID does not exist" containerID="8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.116561 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3"} err="failed to get container status \"8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3\": rpc error: code = NotFound desc = could not find container \"8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3\": container with ID starting with 8146cb52d2fe5d2a5772f1775b4bd0044cc03b4c267f1f42cae0b0a6523374c3 not found: ID does not exist" Oct 01 09:56:47 crc kubenswrapper[4837]: I1001 09:56:47.835452 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" path="/var/lib/kubelet/pods/712274e4-5d3e-4c7b-a779-99fb08d8f5a5/volumes" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.936347 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:56:51 crc kubenswrapper[4837]: E1001 09:56:51.937067 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="extract-content" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.937080 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="extract-content" Oct 01 09:56:51 crc kubenswrapper[4837]: E1001 09:56:51.937102 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="registry-server" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.937108 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="registry-server" Oct 01 09:56:51 crc kubenswrapper[4837]: E1001 09:56:51.937121 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="extract-utilities" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.937127 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="extract-utilities" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.937320 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="712274e4-5d3e-4c7b-a779-99fb08d8f5a5" containerName="registry-server" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.938736 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:51 crc kubenswrapper[4837]: I1001 09:56:51.955575 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.020445 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.020565 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.020683 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfmkm\" (UniqueName: \"kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.121749 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfmkm\" (UniqueName: \"kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.121882 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.121912 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.122318 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.122767 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.149505 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfmkm\" (UniqueName: \"kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm\") pod \"certified-operators-zdb8q\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.263769 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:56:52 crc kubenswrapper[4837]: I1001 09:56:52.795607 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:56:53 crc kubenswrapper[4837]: I1001 09:56:53.038436 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerStarted","Data":"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978"} Oct 01 09:56:53 crc kubenswrapper[4837]: I1001 09:56:53.038488 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerStarted","Data":"3a7142974f40f37b963f62fb90d5855fed89e0747ff8ea7655b53771ba663e85"} Oct 01 09:56:53 crc kubenswrapper[4837]: I1001 09:56:53.079680 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:56:53 crc kubenswrapper[4837]: I1001 09:56:53.079780 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:56:54 crc kubenswrapper[4837]: I1001 09:56:54.052168 4837 generic.go:334] "Generic (PLEG): container finished" podID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerID="9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978" exitCode=0 Oct 01 09:56:54 crc kubenswrapper[4837]: I1001 09:56:54.052230 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerDied","Data":"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978"} Oct 01 09:56:56 crc kubenswrapper[4837]: I1001 09:56:56.086566 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerStarted","Data":"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8"} Oct 01 09:56:57 crc kubenswrapper[4837]: I1001 09:56:57.101877 4837 generic.go:334] "Generic (PLEG): container finished" podID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerID="bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8" exitCode=0 Oct 01 09:56:57 crc kubenswrapper[4837]: I1001 09:56:57.101955 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerDied","Data":"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8"} Oct 01 09:56:59 crc kubenswrapper[4837]: I1001 09:56:59.130870 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerStarted","Data":"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9"} Oct 01 09:56:59 crc kubenswrapper[4837]: I1001 09:56:59.157343 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zdb8q" podStartSLOduration=4.539897893 podStartE2EDuration="8.157318478s" podCreationTimestamp="2025-10-01 09:56:51 +0000 UTC" firstStartedPulling="2025-10-01 09:56:54.055820865 +0000 UTC m=+10270.897428360" lastFinishedPulling="2025-10-01 09:56:57.67324145 +0000 UTC m=+10274.514848945" observedRunningTime="2025-10-01 09:56:59.154945499 +0000 UTC m=+10275.996552964" watchObservedRunningTime="2025-10-01 09:56:59.157318478 +0000 UTC m=+10275.998925963" Oct 01 09:57:02 crc kubenswrapper[4837]: I1001 09:57:02.264198 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:02 crc kubenswrapper[4837]: I1001 09:57:02.264553 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:02 crc kubenswrapper[4837]: I1001 09:57:02.320343 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:03 crc kubenswrapper[4837]: I1001 09:57:03.260753 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:03 crc kubenswrapper[4837]: I1001 09:57:03.574756 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.220925 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zdb8q" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="registry-server" containerID="cri-o://85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9" gracePeriod=2 Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.802824 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.821905 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfmkm\" (UniqueName: \"kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm\") pod \"49ae09c5-2ea1-4660-acbe-831614de1ac8\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.822141 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content\") pod \"49ae09c5-2ea1-4660-acbe-831614de1ac8\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.822198 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities\") pod \"49ae09c5-2ea1-4660-acbe-831614de1ac8\" (UID: \"49ae09c5-2ea1-4660-acbe-831614de1ac8\") " Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.823680 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities" (OuterVolumeSpecName: "utilities") pod "49ae09c5-2ea1-4660-acbe-831614de1ac8" (UID: "49ae09c5-2ea1-4660-acbe-831614de1ac8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.828901 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm" (OuterVolumeSpecName: "kube-api-access-vfmkm") pod "49ae09c5-2ea1-4660-acbe-831614de1ac8" (UID: "49ae09c5-2ea1-4660-acbe-831614de1ac8"). InnerVolumeSpecName "kube-api-access-vfmkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.888906 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49ae09c5-2ea1-4660-acbe-831614de1ac8" (UID: "49ae09c5-2ea1-4660-acbe-831614de1ac8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.927098 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.927171 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49ae09c5-2ea1-4660-acbe-831614de1ac8-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:05 crc kubenswrapper[4837]: I1001 09:57:05.927186 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfmkm\" (UniqueName: \"kubernetes.io/projected/49ae09c5-2ea1-4660-acbe-831614de1ac8-kube-api-access-vfmkm\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.235051 4837 generic.go:334] "Generic (PLEG): container finished" podID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerID="85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9" exitCode=0 Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.235111 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zdb8q" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.235131 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerDied","Data":"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9"} Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.236079 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zdb8q" event={"ID":"49ae09c5-2ea1-4660-acbe-831614de1ac8","Type":"ContainerDied","Data":"3a7142974f40f37b963f62fb90d5855fed89e0747ff8ea7655b53771ba663e85"} Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.236150 4837 scope.go:117] "RemoveContainer" containerID="85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.270442 4837 scope.go:117] "RemoveContainer" containerID="bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.272834 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.283195 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zdb8q"] Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.295780 4837 scope.go:117] "RemoveContainer" containerID="9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.339631 4837 scope.go:117] "RemoveContainer" containerID="85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9" Oct 01 09:57:06 crc kubenswrapper[4837]: E1001 09:57:06.340346 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9\": container with ID starting with 85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9 not found: ID does not exist" containerID="85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.340382 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9"} err="failed to get container status \"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9\": rpc error: code = NotFound desc = could not find container \"85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9\": container with ID starting with 85666a67132d24318cf9cf1ba8bfbd26d92e08cef738e9dbbe90d0ed170047e9 not found: ID does not exist" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.340405 4837 scope.go:117] "RemoveContainer" containerID="bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8" Oct 01 09:57:06 crc kubenswrapper[4837]: E1001 09:57:06.340803 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8\": container with ID starting with bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8 not found: ID does not exist" containerID="bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.340847 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8"} err="failed to get container status \"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8\": rpc error: code = NotFound desc = could not find container \"bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8\": container with ID starting with bbaa805a7cd4e1dfad78845e04d5aa7f8f73c1720beef266f47a5d9c1cf474e8 not found: ID does not exist" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.340877 4837 scope.go:117] "RemoveContainer" containerID="9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978" Oct 01 09:57:06 crc kubenswrapper[4837]: E1001 09:57:06.341258 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978\": container with ID starting with 9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978 not found: ID does not exist" containerID="9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978" Oct 01 09:57:06 crc kubenswrapper[4837]: I1001 09:57:06.341283 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978"} err="failed to get container status \"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978\": rpc error: code = NotFound desc = could not find container \"9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978\": container with ID starting with 9fd8cb8f9eb0c10dc113cb339d2e474f459ecb3500a65d948b64c51fc580f978 not found: ID does not exist" Oct 01 09:57:07 crc kubenswrapper[4837]: I1001 09:57:07.834479 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" path="/var/lib/kubelet/pods/49ae09c5-2ea1-4660-acbe-831614de1ac8/volumes" Oct 01 09:57:17 crc kubenswrapper[4837]: I1001 09:57:17.387274 4837 generic.go:334] "Generic (PLEG): container finished" podID="41ead075-9a6e-4f49-812a-9c8cc8937c22" containerID="2ea25d8e8685cb43bdcfbb936e8600e3508cb41c1a0a3559bd4d850d35be85a8" exitCode=0 Oct 01 09:57:17 crc kubenswrapper[4837]: I1001 09:57:17.387343 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" event={"ID":"41ead075-9a6e-4f49-812a-9c8cc8937c22","Type":"ContainerDied","Data":"2ea25d8e8685cb43bdcfbb936e8600e3508cb41c1a0a3559bd4d850d35be85a8"} Oct 01 09:57:18 crc kubenswrapper[4837]: I1001 09:57:18.931847 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.114976 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key\") pod \"41ead075-9a6e-4f49-812a-9c8cc8937c22\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.115059 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle\") pod \"41ead075-9a6e-4f49-812a-9c8cc8937c22\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.115251 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0\") pod \"41ead075-9a6e-4f49-812a-9c8cc8937c22\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.115287 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory\") pod \"41ead075-9a6e-4f49-812a-9c8cc8937c22\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.115330 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzrcc\" (UniqueName: \"kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc\") pod \"41ead075-9a6e-4f49-812a-9c8cc8937c22\" (UID: \"41ead075-9a6e-4f49-812a-9c8cc8937c22\") " Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.413332 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" event={"ID":"41ead075-9a6e-4f49-812a-9c8cc8937c22","Type":"ContainerDied","Data":"d1dabf3d4554758f574b251016374228ecfe1b88334d43c8687d699ef5b22635"} Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.413748 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1dabf3d4554758f574b251016374228ecfe1b88334d43c8687d699ef5b22635" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.413421 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c4mx" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.649739 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "41ead075-9a6e-4f49-812a-9c8cc8937c22" (UID: "41ead075-9a6e-4f49-812a-9c8cc8937c22"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.650498 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc" (OuterVolumeSpecName: "kube-api-access-rzrcc") pod "41ead075-9a6e-4f49-812a-9c8cc8937c22" (UID: "41ead075-9a6e-4f49-812a-9c8cc8937c22"). InnerVolumeSpecName "kube-api-access-rzrcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.681850 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory" (OuterVolumeSpecName: "inventory") pod "41ead075-9a6e-4f49-812a-9c8cc8937c22" (UID: "41ead075-9a6e-4f49-812a-9c8cc8937c22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.704975 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "41ead075-9a6e-4f49-812a-9c8cc8937c22" (UID: "41ead075-9a6e-4f49-812a-9c8cc8937c22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.709059 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "41ead075-9a6e-4f49-812a-9c8cc8937c22" (UID: "41ead075-9a6e-4f49-812a-9c8cc8937c22"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.730478 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.730822 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.730848 4837 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.730865 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41ead075-9a6e-4f49-812a-9c8cc8937c22-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:19 crc kubenswrapper[4837]: I1001 09:57:19.730879 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzrcc\" (UniqueName: \"kubernetes.io/projected/41ead075-9a6e-4f49-812a-9c8cc8937c22-kube-api-access-rzrcc\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.078953 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.079464 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.079504 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.080274 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.080325 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157" gracePeriod=600 Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.474241 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157" exitCode=0 Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.474347 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157"} Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.474603 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba"} Oct 01 09:57:23 crc kubenswrapper[4837]: I1001 09:57:23.474640 4837 scope.go:117] "RemoveContainer" containerID="fffef18435a15e5c4a17a6262e51249b9b3d1787ee3b7eb97e370dcafc041394" Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.106795 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.107460 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="94df4b1a-2b87-4889-9713-fcc96cbfd579" containerName="nova-cell0-conductor-conductor" containerID="cri-o://3f219857e39bd485a4069aec9a14316e135d4dc2cd3edc38b0d094d07f572d5c" gracePeriod=30 Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.121190 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.121382 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" containerName="nova-cell1-conductor-conductor" containerID="cri-o://0ea8b0ca63d78c0bcc8507109cd8380044e1ffa354eeb56d47eeea2b6a5133b6" gracePeriod=30 Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.841291 4837 generic.go:334] "Generic (PLEG): container finished" podID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" containerID="0ea8b0ca63d78c0bcc8507109cd8380044e1ffa354eeb56d47eeea2b6a5133b6" exitCode=0 Oct 01 09:57:51 crc kubenswrapper[4837]: I1001 09:57:51.841338 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d4b9891-1b5e-42d3-888f-2558f018e4b2","Type":"ContainerDied","Data":"0ea8b0ca63d78c0bcc8507109cd8380044e1ffa354eeb56d47eeea2b6a5133b6"} Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:51.999553 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:51.999943 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" containerID="cri-o://79e2f40b5f6c2e364658462e4dfd2f3a4aadc83ed2cf052842724dcded0d8b05" gracePeriod=30 Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.000385 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" containerID="cri-o://f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2" gracePeriod=30 Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.016518 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.016738 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerName="nova-scheduler-scheduler" containerID="cri-o://e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" gracePeriod=30 Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.048237 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.048588 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" containerID="cri-o://c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f" gracePeriod=30 Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.048871 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" containerID="cri-o://124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826" gracePeriod=30 Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.093742 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5"] Oct 01 09:57:52 crc kubenswrapper[4837]: E1001 09:57:52.094352 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ead075-9a6e-4f49-812a-9c8cc8937c22" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.094426 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ead075-9a6e-4f49-812a-9c8cc8937c22" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 01 09:57:52 crc kubenswrapper[4837]: E1001 09:57:52.094502 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="extract-content" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.094561 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="extract-content" Oct 01 09:57:52 crc kubenswrapper[4837]: E1001 09:57:52.094628 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="extract-utilities" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.094820 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="extract-utilities" Oct 01 09:57:52 crc kubenswrapper[4837]: E1001 09:57:52.094891 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="registry-server" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.094950 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="registry-server" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.095180 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ead075-9a6e-4f49-812a-9c8cc8937c22" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.095248 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="49ae09c5-2ea1-4660-acbe-831614de1ac8" containerName="registry-server" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.096036 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.105875 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.106061 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.106166 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.106261 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.106354 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.106667 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.111076 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-9s5zp" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.112034 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5"] Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.239191 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.267921 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.267965 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.267990 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268127 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268159 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268186 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bltkf\" (UniqueName: \"kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268209 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268393 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.268512 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370024 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k796\" (UniqueName: \"kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796\") pod \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370109 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle\") pod \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370184 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data\") pod \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\" (UID: \"7d4b9891-1b5e-42d3-888f-2558f018e4b2\") " Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370618 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370661 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370705 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bltkf\" (UniqueName: \"kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370725 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.370765 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.371206 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.371325 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.371361 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.371387 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.373283 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.376320 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.376509 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.377480 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.378649 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796" (OuterVolumeSpecName: "kube-api-access-7k796") pod "7d4b9891-1b5e-42d3-888f-2558f018e4b2" (UID: "7d4b9891-1b5e-42d3-888f-2558f018e4b2"). InnerVolumeSpecName "kube-api-access-7k796". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.379410 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.379941 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.381419 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.390083 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.400195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bltkf\" (UniqueName: \"kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.408800 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d4b9891-1b5e-42d3-888f-2558f018e4b2" (UID: "7d4b9891-1b5e-42d3-888f-2558f018e4b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.414973 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data" (OuterVolumeSpecName: "config-data") pod "7d4b9891-1b5e-42d3-888f-2558f018e4b2" (UID: "7d4b9891-1b5e-42d3-888f-2558f018e4b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.473566 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k796\" (UniqueName: \"kubernetes.io/projected/7d4b9891-1b5e-42d3-888f-2558f018e4b2-kube-api-access-7k796\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.473623 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.473649 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d4b9891-1b5e-42d3-888f-2558f018e4b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:52 crc kubenswrapper[4837]: I1001 09:57:52.537644 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.853647 4837 generic.go:334] "Generic (PLEG): container finished" podID="051de007-2b53-46f9-a448-c7019b32d447" containerID="79e2f40b5f6c2e364658462e4dfd2f3a4aadc83ed2cf052842724dcded0d8b05" exitCode=143 Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.853732 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerDied","Data":"79e2f40b5f6c2e364658462e4dfd2f3a4aadc83ed2cf052842724dcded0d8b05"} Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.867051 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7d4b9891-1b5e-42d3-888f-2558f018e4b2","Type":"ContainerDied","Data":"d695359fff4d55a1a3593c5d3d44cdf94e16f74908e7c27a9a59b3b56b5bd923"} Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.867083 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.867134 4837 scope.go:117] "RemoveContainer" containerID="0ea8b0ca63d78c0bcc8507109cd8380044e1ffa354eeb56d47eeea2b6a5133b6" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.869591 4837 generic.go:334] "Generic (PLEG): container finished" podID="47457bb8-964e-452f-886f-9019dd2f818a" containerID="c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f" exitCode=143 Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.869627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerDied","Data":"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f"} Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.923084 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.944112 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.954584 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:53 crc kubenswrapper[4837]: E1001 09:57:52.955108 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" containerName="nova-cell1-conductor-conductor" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.955125 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" containerName="nova-cell1-conductor-conductor" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.955378 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" containerName="nova-cell1-conductor-conductor" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.956334 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.958217 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:52.963351 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.089134 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.089181 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xck2l\" (UniqueName: \"kubernetes.io/projected/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-kube-api-access-xck2l\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.089890 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.191469 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.191558 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.191596 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xck2l\" (UniqueName: \"kubernetes.io/projected/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-kube-api-access-xck2l\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.196341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.196908 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.208756 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xck2l\" (UniqueName: \"kubernetes.io/projected/a532e53d-2421-4f9a-b16e-1bbe87b9a7d7-kube-api-access-xck2l\") pod \"nova-cell1-conductor-0\" (UID: \"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.270772 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.566074 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5"] Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.759212 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:57:53 crc kubenswrapper[4837]: W1001 09:57:53.771654 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda532e53d_2421_4f9a_b16e_1bbe87b9a7d7.slice/crio-7d2447b91d726dc51743d347e716ebd63b5b91c078c815d5f0a939abf33314ca WatchSource:0}: Error finding container 7d2447b91d726dc51743d347e716ebd63b5b91c078c815d5f0a939abf33314ca: Status 404 returned error can't find the container with id 7d2447b91d726dc51743d347e716ebd63b5b91c078c815d5f0a939abf33314ca Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.858141 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4b9891-1b5e-42d3-888f-2558f018e4b2" path="/var/lib/kubelet/pods/7d4b9891-1b5e-42d3-888f-2558f018e4b2/volumes" Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.890174 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" event={"ID":"956da144-c603-4569-b09e-0f4bce2d6c15","Type":"ContainerStarted","Data":"3d720716e7773f0569ce6337528ab90201fb3da7ca9102da1859e61e8d98c716"} Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.897159 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7","Type":"ContainerStarted","Data":"7d2447b91d726dc51743d347e716ebd63b5b91c078c815d5f0a939abf33314ca"} Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.900799 4837 generic.go:334] "Generic (PLEG): container finished" podID="94df4b1a-2b87-4889-9713-fcc96cbfd579" containerID="3f219857e39bd485a4069aec9a14316e135d4dc2cd3edc38b0d094d07f572d5c" exitCode=0 Oct 01 09:57:53 crc kubenswrapper[4837]: I1001 09:57:53.900846 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94df4b1a-2b87-4889-9713-fcc96cbfd579","Type":"ContainerDied","Data":"3f219857e39bd485a4069aec9a14316e135d4dc2cd3edc38b0d094d07f572d5c"} Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.144056 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.321169 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data\") pod \"94df4b1a-2b87-4889-9713-fcc96cbfd579\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.321223 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle\") pod \"94df4b1a-2b87-4889-9713-fcc96cbfd579\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.321423 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st5jn\" (UniqueName: \"kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn\") pod \"94df4b1a-2b87-4889-9713-fcc96cbfd579\" (UID: \"94df4b1a-2b87-4889-9713-fcc96cbfd579\") " Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.325916 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn" (OuterVolumeSpecName: "kube-api-access-st5jn") pod "94df4b1a-2b87-4889-9713-fcc96cbfd579" (UID: "94df4b1a-2b87-4889-9713-fcc96cbfd579"). InnerVolumeSpecName "kube-api-access-st5jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.353835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data" (OuterVolumeSpecName: "config-data") pod "94df4b1a-2b87-4889-9713-fcc96cbfd579" (UID: "94df4b1a-2b87-4889-9713-fcc96cbfd579"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.355858 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94df4b1a-2b87-4889-9713-fcc96cbfd579" (UID: "94df4b1a-2b87-4889-9713-fcc96cbfd579"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.424059 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st5jn\" (UniqueName: \"kubernetes.io/projected/94df4b1a-2b87-4889-9713-fcc96cbfd579-kube-api-access-st5jn\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.424096 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.424106 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94df4b1a-2b87-4889-9713-fcc96cbfd579-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.916410 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"94df4b1a-2b87-4889-9713-fcc96cbfd579","Type":"ContainerDied","Data":"4a705e1ab5bce2bd3d21fa8c5850415034ddeefc00447576f9f1338e2226b501"} Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.916880 4837 scope.go:117] "RemoveContainer" containerID="3f219857e39bd485a4069aec9a14316e135d4dc2cd3edc38b0d094d07f572d5c" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.916514 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.919927 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" event={"ID":"956da144-c603-4569-b09e-0f4bce2d6c15","Type":"ContainerStarted","Data":"e00e23bd8a75be3c810c7a228f35a21e10af01dcec70db87737b63175b4754bd"} Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.922599 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a532e53d-2421-4f9a-b16e-1bbe87b9a7d7","Type":"ContainerStarted","Data":"5814ecb33dc14822e946f67cab67e12cdc3122b19e915afaa4e2ed355b240893"} Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.923584 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 09:57:54 crc kubenswrapper[4837]: I1001 09:57:54.972026 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" podStartSLOduration=2.425411312 podStartE2EDuration="2.971962024s" podCreationTimestamp="2025-10-01 09:57:52 +0000 UTC" firstStartedPulling="2025-10-01 09:57:53.570062265 +0000 UTC m=+10330.411669720" lastFinishedPulling="2025-10-01 09:57:54.116612967 +0000 UTC m=+10330.958220432" observedRunningTime="2025-10-01 09:57:54.950493924 +0000 UTC m=+10331.792101419" watchObservedRunningTime="2025-10-01 09:57:54.971962024 +0000 UTC m=+10331.813569509" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.031813 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.031785511 podStartE2EDuration="3.031785511s" podCreationTimestamp="2025-10-01 09:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:57:54.978875074 +0000 UTC m=+10331.820482569" watchObservedRunningTime="2025-10-01 09:57:55.031785511 +0000 UTC m=+10331.873392976" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.045935 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.056324 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.066168 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.066709 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94df4b1a-2b87-4889-9713-fcc96cbfd579" containerName="nova-cell0-conductor-conductor" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.066734 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="94df4b1a-2b87-4889-9713-fcc96cbfd579" containerName="nova-cell0-conductor-conductor" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.067149 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="94df4b1a-2b87-4889-9713-fcc96cbfd579" containerName="nova-cell0-conductor-conductor" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.068121 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.071384 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.073729 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.199910 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.106:8775/\": read tcp 10.217.0.2:38302->10.217.1.106:8775: read: connection reset by peer" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.200452 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.106:8775/\": read tcp 10.217.0.2:38314->10.217.1.106:8775: read: connection reset by peer" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.242808 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.242882 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.242963 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54rlp\" (UniqueName: \"kubernetes.io/projected/7a7c8913-69d5-47a0-82c1-6136fc655ba6-kube-api-access-54rlp\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.345180 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.345299 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.345409 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54rlp\" (UniqueName: \"kubernetes.io/projected/7a7c8913-69d5-47a0-82c1-6136fc655ba6-kube-api-access-54rlp\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.354930 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.361847 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a7c8913-69d5-47a0-82c1-6136fc655ba6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.365380 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54rlp\" (UniqueName: \"kubernetes.io/projected/7a7c8913-69d5-47a0-82c1-6136fc655ba6-kube-api-access-54rlp\") pod \"nova-cell0-conductor-0\" (UID: \"7a7c8913-69d5-47a0-82c1-6136fc655ba6\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.389058 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.414597 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.109:8774/\": read tcp 10.217.0.2:52798->10.217.1.109:8774: read: connection reset by peer" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.414666 4837 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.109:8774/\": read tcp 10.217.0.2:52788->10.217.1.109:8774: read: connection reset by peer" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.629060 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.769167 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs\") pod \"47457bb8-964e-452f-886f-9019dd2f818a\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.769227 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data\") pod \"47457bb8-964e-452f-886f-9019dd2f818a\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.769343 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle\") pod \"47457bb8-964e-452f-886f-9019dd2f818a\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.769392 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs\") pod \"47457bb8-964e-452f-886f-9019dd2f818a\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.769474 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbbzw\" (UniqueName: \"kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw\") pod \"47457bb8-964e-452f-886f-9019dd2f818a\" (UID: \"47457bb8-964e-452f-886f-9019dd2f818a\") " Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.771633 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs" (OuterVolumeSpecName: "logs") pod "47457bb8-964e-452f-886f-9019dd2f818a" (UID: "47457bb8-964e-452f-886f-9019dd2f818a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.776860 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw" (OuterVolumeSpecName: "kube-api-access-sbbzw") pod "47457bb8-964e-452f-886f-9019dd2f818a" (UID: "47457bb8-964e-452f-886f-9019dd2f818a"). InnerVolumeSpecName "kube-api-access-sbbzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.784772 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod051de007_2b53_46f9_a448_c7019b32d447.slice/crio-conmon-f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod051de007_2b53_46f9_a448_c7019b32d447.slice/crio-f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.802875 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data" (OuterVolumeSpecName: "config-data") pod "47457bb8-964e-452f-886f-9019dd2f818a" (UID: "47457bb8-964e-452f-886f-9019dd2f818a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.807133 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47457bb8-964e-452f-886f-9019dd2f818a" (UID: "47457bb8-964e-452f-886f-9019dd2f818a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.845491 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94df4b1a-2b87-4889-9713-fcc96cbfd579" path="/var/lib/kubelet/pods/94df4b1a-2b87-4889-9713-fcc96cbfd579/volumes" Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.859064 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.861072 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "47457bb8-964e-452f-886f-9019dd2f818a" (UID: "47457bb8-964e-452f-886f-9019dd2f818a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.868570 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.871492 4837 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:57:55 crc kubenswrapper[4837]: E1001 09:57:55.871631 4837 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerName="nova-scheduler-scheduler" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.872580 4837 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.872601 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.872611 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47457bb8-964e-452f-886f-9019dd2f818a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.872619 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47457bb8-964e-452f-886f-9019dd2f818a-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.872650 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbbzw\" (UniqueName: \"kubernetes.io/projected/47457bb8-964e-452f-886f-9019dd2f818a-kube-api-access-sbbzw\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:55 crc kubenswrapper[4837]: W1001 09:57:55.928265 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a7c8913_69d5_47a0_82c1_6136fc655ba6.slice/crio-43ec5d9d1b254507f7f92076c0bace71d79afd8d4d3f654fb1e550f0086a6afe WatchSource:0}: Error finding container 43ec5d9d1b254507f7f92076c0bace71d79afd8d4d3f654fb1e550f0086a6afe: Status 404 returned error can't find the container with id 43ec5d9d1b254507f7f92076c0bace71d79afd8d4d3f654fb1e550f0086a6afe Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.984667 4837 generic.go:334] "Generic (PLEG): container finished" podID="051de007-2b53-46f9-a448-c7019b32d447" containerID="f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2" exitCode=0 Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.987752 4837 generic.go:334] "Generic (PLEG): container finished" podID="47457bb8-964e-452f-886f-9019dd2f818a" containerID="124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826" exitCode=0 Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.988546 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.996044 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.996082 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerDied","Data":"f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2"} Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.996108 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerDied","Data":"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826"} Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.996123 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47457bb8-964e-452f-886f-9019dd2f818a","Type":"ContainerDied","Data":"50bf945ea109e52943eac336fa48d034a14ff440fe8280e9eed8a4605c365c8a"} Oct 01 09:57:55 crc kubenswrapper[4837]: I1001 09:57:55.996140 4837 scope.go:117] "RemoveContainer" containerID="124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.014265 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.036554 4837 scope.go:117] "RemoveContainer" containerID="c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.043200 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.063532 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.080595 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.081086 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081108 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.081146 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081155 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.081176 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081183 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.081194 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081201 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081425 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-log" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081457 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-metadata" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081486 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="051de007-2b53-46f9-a448-c7019b32d447" containerName="nova-api-api" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.081500 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="47457bb8-964e-452f-886f-9019dd2f818a" containerName="nova-metadata-log" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.082647 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.086159 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.086223 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.101743 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.102824 4837 scope.go:117] "RemoveContainer" containerID="124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826" Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.103987 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826\": container with ID starting with 124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826 not found: ID does not exist" containerID="124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.104053 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826"} err="failed to get container status \"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826\": rpc error: code = NotFound desc = could not find container \"124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826\": container with ID starting with 124173529b956dbe829ff719b30a1e1ce2c3ef4af36d811e63b9c60add54e826 not found: ID does not exist" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.104076 4837 scope.go:117] "RemoveContainer" containerID="c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f" Oct 01 09:57:56 crc kubenswrapper[4837]: E1001 09:57:56.104539 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f\": container with ID starting with c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f not found: ID does not exist" containerID="c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.104556 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f"} err="failed to get container status \"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f\": rpc error: code = NotFound desc = could not find container \"c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f\": container with ID starting with c41d248d935aed3b32c10dedc459765bd8b679725e72cf64daca6054c194731f not found: ID does not exist" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185276 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hnm7\" (UniqueName: \"kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185330 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185380 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185435 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185487 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185638 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data\") pod \"051de007-2b53-46f9-a448-c7019b32d447\" (UID: \"051de007-2b53-46f9-a448-c7019b32d447\") " Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.185978 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.186031 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-config-data\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.186063 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db57ec59-8011-46c9-b1ba-e113bb6b8137-logs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.186177 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kcrf\" (UniqueName: \"kubernetes.io/projected/db57ec59-8011-46c9-b1ba-e113bb6b8137-kube-api-access-9kcrf\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.186224 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.186276 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs" (OuterVolumeSpecName: "logs") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.189527 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7" (OuterVolumeSpecName: "kube-api-access-5hnm7") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "kube-api-access-5hnm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.224040 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data" (OuterVolumeSpecName: "config-data") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.226772 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.240933 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.251867 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "051de007-2b53-46f9-a448-c7019b32d447" (UID: "051de007-2b53-46f9-a448-c7019b32d447"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288237 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kcrf\" (UniqueName: \"kubernetes.io/projected/db57ec59-8011-46c9-b1ba-e113bb6b8137-kube-api-access-9kcrf\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288312 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288332 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-config-data\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288391 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db57ec59-8011-46c9-b1ba-e113bb6b8137-logs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288503 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288514 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288523 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hnm7\" (UniqueName: \"kubernetes.io/projected/051de007-2b53-46f9-a448-c7019b32d447-kube-api-access-5hnm7\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288533 4837 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051de007-2b53-46f9-a448-c7019b32d447-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288540 4837 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288549 4837 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/051de007-2b53-46f9-a448-c7019b32d447-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.288860 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db57ec59-8011-46c9-b1ba-e113bb6b8137-logs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.292398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-config-data\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.293251 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.297092 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db57ec59-8011-46c9-b1ba-e113bb6b8137-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.305054 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kcrf\" (UniqueName: \"kubernetes.io/projected/db57ec59-8011-46c9-b1ba-e113bb6b8137-kube-api-access-9kcrf\") pod \"nova-metadata-0\" (UID: \"db57ec59-8011-46c9-b1ba-e113bb6b8137\") " pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.411978 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:57:56 crc kubenswrapper[4837]: I1001 09:57:56.916061 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.000054 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7a7c8913-69d5-47a0-82c1-6136fc655ba6","Type":"ContainerStarted","Data":"15253609a1c468ec7f4ece18756ee7da8575ccdb76f78f8bef39e0b52c789665"} Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.000585 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7a7c8913-69d5-47a0-82c1-6136fc655ba6","Type":"ContainerStarted","Data":"43ec5d9d1b254507f7f92076c0bace71d79afd8d4d3f654fb1e550f0086a6afe"} Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.000741 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.003105 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db57ec59-8011-46c9-b1ba-e113bb6b8137","Type":"ContainerStarted","Data":"73d448bd16685589d5baa289b0ab64efde5cb561f7580d1050b7e780b66465a8"} Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.007447 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.009885 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"051de007-2b53-46f9-a448-c7019b32d447","Type":"ContainerDied","Data":"8feeaca13a24c80d920f836351ca0430bb33ea168710af63916654deff1ad1e8"} Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.009945 4837 scope.go:117] "RemoveContainer" containerID="f18009527dc6e5175208bc0db3348607e451e179d668d84a60a1f2e5c45142c2" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.029327 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.029304914 podStartE2EDuration="3.029304914s" podCreationTimestamp="2025-10-01 09:57:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:57:57.018699282 +0000 UTC m=+10333.860306737" watchObservedRunningTime="2025-10-01 09:57:57.029304914 +0000 UTC m=+10333.870912369" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.068491 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.100507 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.116228 4837 scope.go:117] "RemoveContainer" containerID="79e2f40b5f6c2e364658462e4dfd2f3a4aadc83ed2cf052842724dcded0d8b05" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.118239 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.126336 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.131752 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.134830 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.134848 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.135074 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211680 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40228c63-32a1-4ea4-9e2f-0849056889e5-logs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211730 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211750 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-config-data\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211804 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211887 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.211922 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7q69\" (UniqueName: \"kubernetes.io/projected/40228c63-32a1-4ea4-9e2f-0849056889e5-kube-api-access-c7q69\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314052 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7q69\" (UniqueName: \"kubernetes.io/projected/40228c63-32a1-4ea4-9e2f-0849056889e5-kube-api-access-c7q69\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314177 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314195 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40228c63-32a1-4ea4-9e2f-0849056889e5-logs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314210 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-config-data\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314266 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.314994 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40228c63-32a1-4ea4-9e2f-0849056889e5-logs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.321516 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.322261 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.323161 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.326472 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40228c63-32a1-4ea4-9e2f-0849056889e5-config-data\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.331644 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7q69\" (UniqueName: \"kubernetes.io/projected/40228c63-32a1-4ea4-9e2f-0849056889e5-kube-api-access-c7q69\") pod \"nova-api-0\" (UID: \"40228c63-32a1-4ea4-9e2f-0849056889e5\") " pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.457021 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.827409 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="051de007-2b53-46f9-a448-c7019b32d447" path="/var/lib/kubelet/pods/051de007-2b53-46f9-a448-c7019b32d447/volumes" Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.828678 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47457bb8-964e-452f-886f-9019dd2f818a" path="/var/lib/kubelet/pods/47457bb8-964e-452f-886f-9019dd2f818a/volumes" Oct 01 09:57:57 crc kubenswrapper[4837]: W1001 09:57:57.906827 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40228c63_32a1_4ea4_9e2f_0849056889e5.slice/crio-4c913505bbbf09ba03ce9c19a6c36d787234b40df74cf7cbcc7753286a2451ca WatchSource:0}: Error finding container 4c913505bbbf09ba03ce9c19a6c36d787234b40df74cf7cbcc7753286a2451ca: Status 404 returned error can't find the container with id 4c913505bbbf09ba03ce9c19a6c36d787234b40df74cf7cbcc7753286a2451ca Oct 01 09:57:57 crc kubenswrapper[4837]: I1001 09:57:57.914332 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:57:58 crc kubenswrapper[4837]: I1001 09:57:58.017395 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40228c63-32a1-4ea4-9e2f-0849056889e5","Type":"ContainerStarted","Data":"4c913505bbbf09ba03ce9c19a6c36d787234b40df74cf7cbcc7753286a2451ca"} Oct 01 09:57:58 crc kubenswrapper[4837]: I1001 09:57:58.021019 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db57ec59-8011-46c9-b1ba-e113bb6b8137","Type":"ContainerStarted","Data":"5aff09746009e084d6e71aaadb0d3e05b54940a64f41b1563d6e7a69c30f4f75"} Oct 01 09:57:58 crc kubenswrapper[4837]: I1001 09:57:58.021043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db57ec59-8011-46c9-b1ba-e113bb6b8137","Type":"ContainerStarted","Data":"1163a92fd5de21afc4dd39608f5a9cffc10e2d0a9c99b8a7277eccfbb7c34279"} Oct 01 09:57:59 crc kubenswrapper[4837]: I1001 09:57:59.033916 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40228c63-32a1-4ea4-9e2f-0849056889e5","Type":"ContainerStarted","Data":"87aa0874841097d1051d12423c3b6c6bb5016ef58da100550a73f625fe4ae5da"} Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.045910 4837 generic.go:334] "Generic (PLEG): container finished" podID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerID="e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" exitCode=0 Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.046102 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6","Type":"ContainerDied","Data":"e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e"} Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.048866 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40228c63-32a1-4ea4-9e2f-0849056889e5","Type":"ContainerStarted","Data":"3cd632395892c3a8cc4072e61d6b6089ff018ee321a9104691434fe6ff5847d5"} Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.070301 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.070279398 podStartE2EDuration="4.070279398s" podCreationTimestamp="2025-10-01 09:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:57:58.040902017 +0000 UTC m=+10334.882509472" watchObservedRunningTime="2025-10-01 09:58:00.070279398 +0000 UTC m=+10336.911886863" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.083390 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.083369611 podStartE2EDuration="3.083369611s" podCreationTimestamp="2025-10-01 09:57:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:58:00.068657407 +0000 UTC m=+10336.910264922" watchObservedRunningTime="2025-10-01 09:58:00.083369611 +0000 UTC m=+10336.924977086" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.146972 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.289019 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7hnz\" (UniqueName: \"kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz\") pod \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.289256 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data\") pod \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.289340 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle\") pod \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\" (UID: \"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6\") " Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.295802 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz" (OuterVolumeSpecName: "kube-api-access-m7hnz") pod "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" (UID: "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6"). InnerVolumeSpecName "kube-api-access-m7hnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.321905 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" (UID: "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.324158 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data" (OuterVolumeSpecName: "config-data") pod "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" (UID: "1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.391837 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.392004 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7hnz\" (UniqueName: \"kubernetes.io/projected/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-kube-api-access-m7hnz\") on node \"crc\" DevicePath \"\"" Oct 01 09:58:00 crc kubenswrapper[4837]: I1001 09:58:00.392083 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.063482 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.063473 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6","Type":"ContainerDied","Data":"5d8fbbf64463382b95030bfdb0374eb92b780dc108b780a5082847332cff5003"} Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.063572 4837 scope.go:117] "RemoveContainer" containerID="e1f280c37c135f2d52c66dea4a94364f4faa1011994ed01147390873a067730e" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.148508 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.166336 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.178233 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:58:01 crc kubenswrapper[4837]: E1001 09:58:01.178988 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerName="nova-scheduler-scheduler" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.179020 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerName="nova-scheduler-scheduler" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.179428 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" containerName="nova-scheduler-scheduler" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.180771 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.184901 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.192652 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.315192 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.315562 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-config-data\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.315825 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpcpc\" (UniqueName: \"kubernetes.io/projected/7fcdf164-8ca8-44f8-b44b-763300203256-kube-api-access-dpcpc\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.412948 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.412998 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.418545 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-config-data\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.418807 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpcpc\" (UniqueName: \"kubernetes.io/projected/7fcdf164-8ca8-44f8-b44b-763300203256-kube-api-access-dpcpc\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.419101 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.426079 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.430742 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fcdf164-8ca8-44f8-b44b-763300203256-config-data\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.446257 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpcpc\" (UniqueName: \"kubernetes.io/projected/7fcdf164-8ca8-44f8-b44b-763300203256-kube-api-access-dpcpc\") pod \"nova-scheduler-0\" (UID: \"7fcdf164-8ca8-44f8-b44b-763300203256\") " pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.512497 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:58:01 crc kubenswrapper[4837]: I1001 09:58:01.829172 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6" path="/var/lib/kubelet/pods/1ff7ee1c-b75f-41a4-8e4f-82ad7631acf6/volumes" Oct 01 09:58:02 crc kubenswrapper[4837]: I1001 09:58:02.053163 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:58:02 crc kubenswrapper[4837]: W1001 09:58:02.073477 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fcdf164_8ca8_44f8_b44b_763300203256.slice/crio-cc2583effcc6ef2e6a817aa1f8bb77f57632674a68df4ca08481048837cd82b9 WatchSource:0}: Error finding container cc2583effcc6ef2e6a817aa1f8bb77f57632674a68df4ca08481048837cd82b9: Status 404 returned error can't find the container with id cc2583effcc6ef2e6a817aa1f8bb77f57632674a68df4ca08481048837cd82b9 Oct 01 09:58:03 crc kubenswrapper[4837]: I1001 09:58:03.096627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fcdf164-8ca8-44f8-b44b-763300203256","Type":"ContainerStarted","Data":"51f64ac5a0c86262b64ddecbb81c4bb62504047328ed0eae2a5d26b18e4396b6"} Oct 01 09:58:03 crc kubenswrapper[4837]: I1001 09:58:03.097071 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fcdf164-8ca8-44f8-b44b-763300203256","Type":"ContainerStarted","Data":"cc2583effcc6ef2e6a817aa1f8bb77f57632674a68df4ca08481048837cd82b9"} Oct 01 09:58:03 crc kubenswrapper[4837]: I1001 09:58:03.116747 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.116726696 podStartE2EDuration="2.116726696s" podCreationTimestamp="2025-10-01 09:58:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:58:03.112627675 +0000 UTC m=+10339.954235130" watchObservedRunningTime="2025-10-01 09:58:03.116726696 +0000 UTC m=+10339.958334151" Oct 01 09:58:03 crc kubenswrapper[4837]: I1001 09:58:03.307242 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 09:58:05 crc kubenswrapper[4837]: I1001 09:58:05.431341 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 09:58:06 crc kubenswrapper[4837]: I1001 09:58:06.413341 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:58:06 crc kubenswrapper[4837]: I1001 09:58:06.414113 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:58:06 crc kubenswrapper[4837]: I1001 09:58:06.513092 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:58:07 crc kubenswrapper[4837]: I1001 09:58:07.430228 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="db57ec59-8011-46c9-b1ba-e113bb6b8137" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:58:07 crc kubenswrapper[4837]: I1001 09:58:07.430909 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="db57ec59-8011-46c9-b1ba-e113bb6b8137" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:58:07 crc kubenswrapper[4837]: I1001 09:58:07.458160 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:58:07 crc kubenswrapper[4837]: I1001 09:58:07.458209 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:58:08 crc kubenswrapper[4837]: I1001 09:58:08.474023 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40228c63-32a1-4ea4-9e2f-0849056889e5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:58:08 crc kubenswrapper[4837]: I1001 09:58:08.474011 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40228c63-32a1-4ea4-9e2f-0849056889e5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:58:11 crc kubenswrapper[4837]: I1001 09:58:11.513368 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 09:58:11 crc kubenswrapper[4837]: I1001 09:58:11.558301 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 09:58:12 crc kubenswrapper[4837]: I1001 09:58:12.224741 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 09:58:16 crc kubenswrapper[4837]: I1001 09:58:16.418062 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:58:16 crc kubenswrapper[4837]: I1001 09:58:16.418757 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:58:16 crc kubenswrapper[4837]: I1001 09:58:16.422026 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:58:16 crc kubenswrapper[4837]: I1001 09:58:16.422869 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:58:17 crc kubenswrapper[4837]: I1001 09:58:17.466485 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:58:17 crc kubenswrapper[4837]: I1001 09:58:17.467403 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:58:17 crc kubenswrapper[4837]: I1001 09:58:17.472750 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:58:17 crc kubenswrapper[4837]: I1001 09:58:17.474302 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:58:18 crc kubenswrapper[4837]: I1001 09:58:18.283058 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:58:18 crc kubenswrapper[4837]: I1001 09:58:18.294470 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:59:23 crc kubenswrapper[4837]: I1001 09:59:23.079433 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:59:23 crc kubenswrapper[4837]: I1001 09:59:23.080246 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:59:53 crc kubenswrapper[4837]: I1001 09:59:53.079120 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:59:53 crc kubenswrapper[4837]: I1001 09:59:53.079755 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.655835 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.662159 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.667071 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.751881 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.752479 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.752575 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjt77\" (UniqueName: \"kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.855415 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.855978 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.856310 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.857568 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.856182 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjt77\" (UniqueName: \"kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.879499 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjt77\" (UniqueName: \"kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77\") pod \"redhat-marketplace-5swcw\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:56 crc kubenswrapper[4837]: I1001 09:59:56.988341 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 09:59:57 crc kubenswrapper[4837]: I1001 09:59:57.460286 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 09:59:57 crc kubenswrapper[4837]: I1001 09:59:57.543939 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerStarted","Data":"b3a3866bdce3b3737c7ca97d697e4bf2fdbf31292ee0b493aac46e5ab2a6e0c2"} Oct 01 09:59:58 crc kubenswrapper[4837]: I1001 09:59:58.563302 4837 generic.go:334] "Generic (PLEG): container finished" podID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerID="85d7121d3a012da5c6b778ea4ca2ff2d54761bed5515a9b438ac0961f05b553e" exitCode=0 Oct 01 09:59:58 crc kubenswrapper[4837]: I1001 09:59:58.563868 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerDied","Data":"85d7121d3a012da5c6b778ea4ca2ff2d54761bed5515a9b438ac0961f05b553e"} Oct 01 09:59:59 crc kubenswrapper[4837]: I1001 09:59:59.585154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerStarted","Data":"62b720018f4e924dc541a74e11d5644c14aa5ce9bb72a58de710f9998a605d6a"} Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.167369 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl"] Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.170810 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.175279 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.176287 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.178061 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl"] Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.332731 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.333589 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjbvd\" (UniqueName: \"kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.333768 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.436552 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.436876 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjbvd\" (UniqueName: \"kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.436910 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.438376 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.448743 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.461976 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjbvd\" (UniqueName: \"kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd\") pod \"collect-profiles-29321880-6rsjl\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:00 crc kubenswrapper[4837]: I1001 10:00:00.498476 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.018930 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl"] Oct 01 10:00:01 crc kubenswrapper[4837]: W1001 10:00:01.023863 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod796d73ac_53c6_4359_bdb9_a311ef6a597e.slice/crio-1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53 WatchSource:0}: Error finding container 1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53: Status 404 returned error can't find the container with id 1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53 Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.614142 4837 generic.go:334] "Generic (PLEG): container finished" podID="796d73ac-53c6-4359-bdb9-a311ef6a597e" containerID="5a14feb6a805582b9e1a76e670633d8008329f567ec75e4257ff0ac5bca98bf8" exitCode=0 Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.614223 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" event={"ID":"796d73ac-53c6-4359-bdb9-a311ef6a597e","Type":"ContainerDied","Data":"5a14feb6a805582b9e1a76e670633d8008329f567ec75e4257ff0ac5bca98bf8"} Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.614707 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" event={"ID":"796d73ac-53c6-4359-bdb9-a311ef6a597e","Type":"ContainerStarted","Data":"1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53"} Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.617103 4837 generic.go:334] "Generic (PLEG): container finished" podID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerID="62b720018f4e924dc541a74e11d5644c14aa5ce9bb72a58de710f9998a605d6a" exitCode=0 Oct 01 10:00:01 crc kubenswrapper[4837]: I1001 10:00:01.617154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerDied","Data":"62b720018f4e924dc541a74e11d5644c14aa5ce9bb72a58de710f9998a605d6a"} Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.025923 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.205658 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume\") pod \"796d73ac-53c6-4359-bdb9-a311ef6a597e\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.205753 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjbvd\" (UniqueName: \"kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd\") pod \"796d73ac-53c6-4359-bdb9-a311ef6a597e\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.205861 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume\") pod \"796d73ac-53c6-4359-bdb9-a311ef6a597e\" (UID: \"796d73ac-53c6-4359-bdb9-a311ef6a597e\") " Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.206626 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume" (OuterVolumeSpecName: "config-volume") pod "796d73ac-53c6-4359-bdb9-a311ef6a597e" (UID: "796d73ac-53c6-4359-bdb9-a311ef6a597e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.212639 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "796d73ac-53c6-4359-bdb9-a311ef6a597e" (UID: "796d73ac-53c6-4359-bdb9-a311ef6a597e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.213722 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd" (OuterVolumeSpecName: "kube-api-access-wjbvd") pod "796d73ac-53c6-4359-bdb9-a311ef6a597e" (UID: "796d73ac-53c6-4359-bdb9-a311ef6a597e"). InnerVolumeSpecName "kube-api-access-wjbvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.308380 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/796d73ac-53c6-4359-bdb9-a311ef6a597e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.308802 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjbvd\" (UniqueName: \"kubernetes.io/projected/796d73ac-53c6-4359-bdb9-a311ef6a597e-kube-api-access-wjbvd\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.308812 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/796d73ac-53c6-4359-bdb9-a311ef6a597e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.648669 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.648658 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl" event={"ID":"796d73ac-53c6-4359-bdb9-a311ef6a597e","Type":"ContainerDied","Data":"1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53"} Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.648831 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d2a34d929276799a625e9d65d6b37eeae290df39f0941320d31e51abda1df53" Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.652712 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerStarted","Data":"4043cc2583fbbfaa473398c163a80f2847d49053ae389a876b10fcf78adb0bb5"} Oct 01 10:00:03 crc kubenswrapper[4837]: I1001 10:00:03.684829 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5swcw" podStartSLOduration=3.725277035 podStartE2EDuration="7.684809156s" podCreationTimestamp="2025-10-01 09:59:56 +0000 UTC" firstStartedPulling="2025-10-01 09:59:58.568526868 +0000 UTC m=+10455.410134353" lastFinishedPulling="2025-10-01 10:00:02.528058989 +0000 UTC m=+10459.369666474" observedRunningTime="2025-10-01 10:00:03.675129437 +0000 UTC m=+10460.516736892" watchObservedRunningTime="2025-10-01 10:00:03.684809156 +0000 UTC m=+10460.526416621" Oct 01 10:00:04 crc kubenswrapper[4837]: I1001 10:00:04.122013 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9"] Oct 01 10:00:04 crc kubenswrapper[4837]: I1001 10:00:04.131834 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321835-rp4m9"] Oct 01 10:00:05 crc kubenswrapper[4837]: I1001 10:00:05.837619 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65feda55-7457-461d-abd9-a80d76d421c3" path="/var/lib/kubelet/pods/65feda55-7457-461d-abd9-a80d76d421c3/volumes" Oct 01 10:00:06 crc kubenswrapper[4837]: I1001 10:00:06.989913 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:06 crc kubenswrapper[4837]: I1001 10:00:06.990318 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:07 crc kubenswrapper[4837]: I1001 10:00:07.057982 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:07 crc kubenswrapper[4837]: I1001 10:00:07.778855 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:07 crc kubenswrapper[4837]: I1001 10:00:07.857554 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 10:00:09 crc kubenswrapper[4837]: I1001 10:00:09.729593 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5swcw" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="registry-server" containerID="cri-o://4043cc2583fbbfaa473398c163a80f2847d49053ae389a876b10fcf78adb0bb5" gracePeriod=2 Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.749745 4837 generic.go:334] "Generic (PLEG): container finished" podID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerID="4043cc2583fbbfaa473398c163a80f2847d49053ae389a876b10fcf78adb0bb5" exitCode=0 Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.749968 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerDied","Data":"4043cc2583fbbfaa473398c163a80f2847d49053ae389a876b10fcf78adb0bb5"} Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.750211 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5swcw" event={"ID":"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a","Type":"ContainerDied","Data":"b3a3866bdce3b3737c7ca97d697e4bf2fdbf31292ee0b493aac46e5ab2a6e0c2"} Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.750229 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3a3866bdce3b3737c7ca97d697e4bf2fdbf31292ee0b493aac46e5ab2a6e0c2" Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.802078 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.916761 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content\") pod \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.916998 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities\") pod \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.917146 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjt77\" (UniqueName: \"kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77\") pod \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\" (UID: \"78cfccc8-b7b3-4800-bd3a-54c79a45ca0a\") " Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.920422 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities" (OuterVolumeSpecName: "utilities") pod "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" (UID: "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.928582 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77" (OuterVolumeSpecName: "kube-api-access-kjt77") pod "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" (UID: "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a"). InnerVolumeSpecName "kube-api-access-kjt77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:00:10 crc kubenswrapper[4837]: I1001 10:00:10.938653 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" (UID: "78cfccc8-b7b3-4800-bd3a-54c79a45ca0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.020151 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.020501 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.020566 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjt77\" (UniqueName: \"kubernetes.io/projected/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a-kube-api-access-kjt77\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.763259 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5swcw" Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.812827 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 10:00:11 crc kubenswrapper[4837]: I1001 10:00:11.832997 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5swcw"] Oct 01 10:00:13 crc kubenswrapper[4837]: I1001 10:00:13.829988 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" path="/var/lib/kubelet/pods/78cfccc8-b7b3-4800-bd3a-54c79a45ca0a/volumes" Oct 01 10:00:16 crc kubenswrapper[4837]: I1001 10:00:16.776354 4837 scope.go:117] "RemoveContainer" containerID="a4462a7845d32476f3a0f3230099099a685b1d0e8f7bdde77d5c993ccfc5190c" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.079565 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.080379 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.080479 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.082124 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.082299 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" gracePeriod=600 Oct 01 10:00:23 crc kubenswrapper[4837]: E1001 10:00:23.581552 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.984600 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" exitCode=0 Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.984658 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba"} Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.984756 4837 scope.go:117] "RemoveContainer" containerID="541164c0e41d48d4c17d666ced591ebc2aac5dd09e6fdb31700dace7d35bc157" Oct 01 10:00:23 crc kubenswrapper[4837]: I1001 10:00:23.985665 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:00:23 crc kubenswrapper[4837]: E1001 10:00:23.986047 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:00:38 crc kubenswrapper[4837]: I1001 10:00:38.817847 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:00:38 crc kubenswrapper[4837]: E1001 10:00:38.818615 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:00:52 crc kubenswrapper[4837]: I1001 10:00:52.816478 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:00:52 crc kubenswrapper[4837]: E1001 10:00:52.817258 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.161516 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29321881-q6spk"] Oct 01 10:01:00 crc kubenswrapper[4837]: E1001 10:01:00.162708 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.162730 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4837]: E1001 10:01:00.162749 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796d73ac-53c6-4359-bdb9-a311ef6a597e" containerName="collect-profiles" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.162757 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="796d73ac-53c6-4359-bdb9-a311ef6a597e" containerName="collect-profiles" Oct 01 10:01:00 crc kubenswrapper[4837]: E1001 10:01:00.162771 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="extract-content" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.162779 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="extract-content" Oct 01 10:01:00 crc kubenswrapper[4837]: E1001 10:01:00.162799 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="extract-utilities" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.162806 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="extract-utilities" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.163072 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="796d73ac-53c6-4359-bdb9-a311ef6a597e" containerName="collect-profiles" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.163098 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cfccc8-b7b3-4800-bd3a-54c79a45ca0a" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.163980 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.175469 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321881-q6spk"] Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.292034 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.292112 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.293064 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j72s\" (UniqueName: \"kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.293144 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.395331 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j72s\" (UniqueName: \"kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.395380 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.395456 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.395485 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.404858 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.407158 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.416804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.426302 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j72s\" (UniqueName: \"kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s\") pod \"keystone-cron-29321881-q6spk\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.488758 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:00 crc kubenswrapper[4837]: I1001 10:01:00.967977 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321881-q6spk"] Oct 01 10:01:01 crc kubenswrapper[4837]: I1001 10:01:01.459541 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-q6spk" event={"ID":"bce0fa96-7165-40ce-b234-74e76950f708","Type":"ContainerStarted","Data":"2ba25974fda70c818f081f3613929346f1064e8de62e077f04e6cb7ebe62a568"} Oct 01 10:01:01 crc kubenswrapper[4837]: I1001 10:01:01.459617 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-q6spk" event={"ID":"bce0fa96-7165-40ce-b234-74e76950f708","Type":"ContainerStarted","Data":"d67a101d4961572795202f7c23614bdc47f67628e2c45cbaf4ffbd7be265d928"} Oct 01 10:01:01 crc kubenswrapper[4837]: I1001 10:01:01.482074 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29321881-q6spk" podStartSLOduration=1.482053051 podStartE2EDuration="1.482053051s" podCreationTimestamp="2025-10-01 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:01:01.480439541 +0000 UTC m=+10518.322047046" watchObservedRunningTime="2025-10-01 10:01:01.482053051 +0000 UTC m=+10518.323660506" Oct 01 10:01:05 crc kubenswrapper[4837]: I1001 10:01:05.820095 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:01:05 crc kubenswrapper[4837]: E1001 10:01:05.821147 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:01:06 crc kubenswrapper[4837]: I1001 10:01:06.522268 4837 generic.go:334] "Generic (PLEG): container finished" podID="bce0fa96-7165-40ce-b234-74e76950f708" containerID="2ba25974fda70c818f081f3613929346f1064e8de62e077f04e6cb7ebe62a568" exitCode=0 Oct 01 10:01:06 crc kubenswrapper[4837]: I1001 10:01:06.522339 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-q6spk" event={"ID":"bce0fa96-7165-40ce-b234-74e76950f708","Type":"ContainerDied","Data":"2ba25974fda70c818f081f3613929346f1064e8de62e077f04e6cb7ebe62a568"} Oct 01 10:01:07 crc kubenswrapper[4837]: I1001 10:01:07.896445 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:07 crc kubenswrapper[4837]: I1001 10:01:07.981524 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle\") pod \"bce0fa96-7165-40ce-b234-74e76950f708\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " Oct 01 10:01:07 crc kubenswrapper[4837]: I1001 10:01:07.981612 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data\") pod \"bce0fa96-7165-40ce-b234-74e76950f708\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " Oct 01 10:01:07 crc kubenswrapper[4837]: I1001 10:01:07.981655 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys\") pod \"bce0fa96-7165-40ce-b234-74e76950f708\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " Oct 01 10:01:07 crc kubenswrapper[4837]: I1001 10:01:07.981797 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j72s\" (UniqueName: \"kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s\") pod \"bce0fa96-7165-40ce-b234-74e76950f708\" (UID: \"bce0fa96-7165-40ce-b234-74e76950f708\") " Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.006200 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bce0fa96-7165-40ce-b234-74e76950f708" (UID: "bce0fa96-7165-40ce-b234-74e76950f708"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.017354 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s" (OuterVolumeSpecName: "kube-api-access-2j72s") pod "bce0fa96-7165-40ce-b234-74e76950f708" (UID: "bce0fa96-7165-40ce-b234-74e76950f708"). InnerVolumeSpecName "kube-api-access-2j72s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.023736 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bce0fa96-7165-40ce-b234-74e76950f708" (UID: "bce0fa96-7165-40ce-b234-74e76950f708"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.045828 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data" (OuterVolumeSpecName: "config-data") pod "bce0fa96-7165-40ce-b234-74e76950f708" (UID: "bce0fa96-7165-40ce-b234-74e76950f708"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.084293 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.084335 4837 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.084349 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j72s\" (UniqueName: \"kubernetes.io/projected/bce0fa96-7165-40ce-b234-74e76950f708-kube-api-access-2j72s\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.084361 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce0fa96-7165-40ce-b234-74e76950f708-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.550481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-q6spk" event={"ID":"bce0fa96-7165-40ce-b234-74e76950f708","Type":"ContainerDied","Data":"d67a101d4961572795202f7c23614bdc47f67628e2c45cbaf4ffbd7be265d928"} Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.550534 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d67a101d4961572795202f7c23614bdc47f67628e2c45cbaf4ffbd7be265d928" Oct 01 10:01:08 crc kubenswrapper[4837]: I1001 10:01:08.550579 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-q6spk" Oct 01 10:01:20 crc kubenswrapper[4837]: I1001 10:01:20.817509 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:01:20 crc kubenswrapper[4837]: E1001 10:01:20.818652 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:01:30 crc kubenswrapper[4837]: I1001 10:01:30.803280 4837 generic.go:334] "Generic (PLEG): container finished" podID="956da144-c603-4569-b09e-0f4bce2d6c15" containerID="e00e23bd8a75be3c810c7a228f35a21e10af01dcec70db87737b63175b4754bd" exitCode=0 Oct 01 10:01:30 crc kubenswrapper[4837]: I1001 10:01:30.803363 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" event={"ID":"956da144-c603-4569-b09e-0f4bce2d6c15","Type":"ContainerDied","Data":"e00e23bd8a75be3c810c7a228f35a21e10af01dcec70db87737b63175b4754bd"} Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.686879 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:31 crc kubenswrapper[4837]: E1001 10:01:31.687814 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce0fa96-7165-40ce-b234-74e76950f708" containerName="keystone-cron" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.687838 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce0fa96-7165-40ce-b234-74e76950f708" containerName="keystone-cron" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.688103 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce0fa96-7165-40ce-b234-74e76950f708" containerName="keystone-cron" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.690151 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.701442 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.813029 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.813103 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.813189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7c4\" (UniqueName: \"kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.915261 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.915714 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7c4\" (UniqueName: \"kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.915833 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.915849 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.916586 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:31 crc kubenswrapper[4837]: I1001 10:01:31.945605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7c4\" (UniqueName: \"kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4\") pod \"redhat-operators-dfnxd\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.021710 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.421119 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.545997 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546058 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546093 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546132 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546221 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546331 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546440 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546510 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bltkf\" (UniqueName: \"kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.546543 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0\") pod \"956da144-c603-4569-b09e-0f4bce2d6c15\" (UID: \"956da144-c603-4569-b09e-0f4bce2d6c15\") " Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.552552 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf" (OuterVolumeSpecName: "kube-api-access-bltkf") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "kube-api-access-bltkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.560217 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.577632 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.586209 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.600672 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.602050 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory" (OuterVolumeSpecName: "inventory") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: W1001 10:01:32.613211 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedd8be8e_3eb3_4a8a_9b5d_60c44a9e98d7.slice/crio-2be781b3f360698128c773046d50ebb14a75fdbcff6b514e1ca766412f497811 WatchSource:0}: Error finding container 2be781b3f360698128c773046d50ebb14a75fdbcff6b514e1ca766412f497811: Status 404 returned error can't find the container with id 2be781b3f360698128c773046d50ebb14a75fdbcff6b514e1ca766412f497811 Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.618096 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.620900 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.622450 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.635213 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "956da144-c603-4569-b09e-0f4bce2d6c15" (UID: "956da144-c603-4569-b09e-0f4bce2d6c15"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648444 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648485 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648494 4837 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648503 4837 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648521 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648531 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648541 4837 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648549 4837 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/956da144-c603-4569-b09e-0f4bce2d6c15-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.648558 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bltkf\" (UniqueName: \"kubernetes.io/projected/956da144-c603-4569-b09e-0f4bce2d6c15-kube-api-access-bltkf\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.826499 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.826502 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5" event={"ID":"956da144-c603-4569-b09e-0f4bce2d6c15","Type":"ContainerDied","Data":"3d720716e7773f0569ce6337528ab90201fb3da7ca9102da1859e61e8d98c716"} Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.827103 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d720716e7773f0569ce6337528ab90201fb3da7ca9102da1859e61e8d98c716" Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.839179 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerStarted","Data":"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96"} Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.839224 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerStarted","Data":"2be781b3f360698128c773046d50ebb14a75fdbcff6b514e1ca766412f497811"} Oct 01 10:01:32 crc kubenswrapper[4837]: I1001 10:01:32.846277 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:01:33 crc kubenswrapper[4837]: I1001 10:01:33.870847 4837 generic.go:334] "Generic (PLEG): container finished" podID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerID="80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96" exitCode=0 Oct 01 10:01:33 crc kubenswrapper[4837]: I1001 10:01:33.873942 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerDied","Data":"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96"} Oct 01 10:01:34 crc kubenswrapper[4837]: I1001 10:01:34.885635 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerStarted","Data":"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990"} Oct 01 10:01:35 crc kubenswrapper[4837]: I1001 10:01:35.816596 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:01:35 crc kubenswrapper[4837]: E1001 10:01:35.817112 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:01:38 crc kubenswrapper[4837]: I1001 10:01:38.938169 4837 generic.go:334] "Generic (PLEG): container finished" podID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerID="bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990" exitCode=0 Oct 01 10:01:38 crc kubenswrapper[4837]: I1001 10:01:38.938218 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerDied","Data":"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990"} Oct 01 10:01:39 crc kubenswrapper[4837]: I1001 10:01:39.950371 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerStarted","Data":"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994"} Oct 01 10:01:39 crc kubenswrapper[4837]: I1001 10:01:39.974872 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dfnxd" podStartSLOduration=2.412420653 podStartE2EDuration="8.974855175s" podCreationTimestamp="2025-10-01 10:01:31 +0000 UTC" firstStartedPulling="2025-10-01 10:01:32.845575187 +0000 UTC m=+10549.687182632" lastFinishedPulling="2025-10-01 10:01:39.408009689 +0000 UTC m=+10556.249617154" observedRunningTime="2025-10-01 10:01:39.967929663 +0000 UTC m=+10556.809537138" watchObservedRunningTime="2025-10-01 10:01:39.974855175 +0000 UTC m=+10556.816462630" Oct 01 10:01:42 crc kubenswrapper[4837]: I1001 10:01:42.022303 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:42 crc kubenswrapper[4837]: I1001 10:01:42.022666 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:43 crc kubenswrapper[4837]: I1001 10:01:43.086348 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dfnxd" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="registry-server" probeResult="failure" output=< Oct 01 10:01:43 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:01:43 crc kubenswrapper[4837]: > Oct 01 10:01:48 crc kubenswrapper[4837]: I1001 10:01:48.815595 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:01:48 crc kubenswrapper[4837]: E1001 10:01:48.816418 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:01:52 crc kubenswrapper[4837]: I1001 10:01:52.105401 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:52 crc kubenswrapper[4837]: I1001 10:01:52.170643 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:52 crc kubenswrapper[4837]: I1001 10:01:52.347545 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.139377 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dfnxd" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="registry-server" containerID="cri-o://8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994" gracePeriod=2 Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.651191 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.742362 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h7c4\" (UniqueName: \"kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4\") pod \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.743714 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities\") pod \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.744107 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content\") pod \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\" (UID: \"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7\") " Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.744985 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities" (OuterVolumeSpecName: "utilities") pod "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" (UID: "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.754137 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4" (OuterVolumeSpecName: "kube-api-access-5h7c4") pod "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" (UID: "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7"). InnerVolumeSpecName "kube-api-access-5h7c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.847624 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h7c4\" (UniqueName: \"kubernetes.io/projected/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-kube-api-access-5h7c4\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.847944 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.851620 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" (UID: "edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:01:54 crc kubenswrapper[4837]: I1001 10:01:54.949496 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.156009 4837 generic.go:334] "Generic (PLEG): container finished" podID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerID="8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994" exitCode=0 Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.156101 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerDied","Data":"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994"} Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.156137 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dfnxd" event={"ID":"edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7","Type":"ContainerDied","Data":"2be781b3f360698128c773046d50ebb14a75fdbcff6b514e1ca766412f497811"} Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.156133 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dfnxd" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.156159 4837 scope.go:117] "RemoveContainer" containerID="8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.208446 4837 scope.go:117] "RemoveContainer" containerID="bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.208987 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.221532 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dfnxd"] Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.674960 4837 scope.go:117] "RemoveContainer" containerID="80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.729536 4837 scope.go:117] "RemoveContainer" containerID="8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994" Oct 01 10:01:55 crc kubenswrapper[4837]: E1001 10:01:55.730036 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994\": container with ID starting with 8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994 not found: ID does not exist" containerID="8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.730078 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994"} err="failed to get container status \"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994\": rpc error: code = NotFound desc = could not find container \"8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994\": container with ID starting with 8d7bed1463d8e681533f7c2a0c7b3493079093cef71b29204186b2c73973b994 not found: ID does not exist" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.730131 4837 scope.go:117] "RemoveContainer" containerID="bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990" Oct 01 10:01:55 crc kubenswrapper[4837]: E1001 10:01:55.730728 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990\": container with ID starting with bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990 not found: ID does not exist" containerID="bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.730786 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990"} err="failed to get container status \"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990\": rpc error: code = NotFound desc = could not find container \"bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990\": container with ID starting with bc39a424616854863d97fd6c522fcf5f05f24e0724f6e197949f6068469fe990 not found: ID does not exist" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.730820 4837 scope.go:117] "RemoveContainer" containerID="80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96" Oct 01 10:01:55 crc kubenswrapper[4837]: E1001 10:01:55.731193 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96\": container with ID starting with 80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96 not found: ID does not exist" containerID="80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.731231 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96"} err="failed to get container status \"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96\": rpc error: code = NotFound desc = could not find container \"80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96\": container with ID starting with 80b10f03705b41fb2a9b26d15fa20b6ca9761e55424c3f768c5f4c93d5326d96 not found: ID does not exist" Oct 01 10:01:55 crc kubenswrapper[4837]: I1001 10:01:55.830717 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" path="/var/lib/kubelet/pods/edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7/volumes" Oct 01 10:02:01 crc kubenswrapper[4837]: I1001 10:02:01.816368 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:02:01 crc kubenswrapper[4837]: E1001 10:02:01.817481 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:02:14 crc kubenswrapper[4837]: I1001 10:02:14.816506 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:02:14 crc kubenswrapper[4837]: E1001 10:02:14.817379 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:02:27 crc kubenswrapper[4837]: I1001 10:02:27.816134 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:02:27 crc kubenswrapper[4837]: E1001 10:02:27.816974 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:02:42 crc kubenswrapper[4837]: I1001 10:02:42.817332 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:02:42 crc kubenswrapper[4837]: E1001 10:02:42.818187 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:02:57 crc kubenswrapper[4837]: I1001 10:02:57.816440 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:02:57 crc kubenswrapper[4837]: E1001 10:02:57.817202 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:03:09 crc kubenswrapper[4837]: I1001 10:03:09.817443 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:03:09 crc kubenswrapper[4837]: E1001 10:03:09.818527 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:03:19 crc kubenswrapper[4837]: I1001 10:03:19.919109 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 10:03:19 crc kubenswrapper[4837]: I1001 10:03:19.919684 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="22907b50-b230-4f94-bd39-7d66e4601e08" containerName="adoption" containerID="cri-o://ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86" gracePeriod=30 Oct 01 10:03:22 crc kubenswrapper[4837]: I1001 10:03:22.815892 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:03:22 crc kubenswrapper[4837]: E1001 10:03:22.816475 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:03:36 crc kubenswrapper[4837]: I1001 10:03:36.816406 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:03:36 crc kubenswrapper[4837]: E1001 10:03:36.817208 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:03:49 crc kubenswrapper[4837]: I1001 10:03:49.815955 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:03:49 crc kubenswrapper[4837]: E1001 10:03:49.816793 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.441826 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.461730 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") pod \"22907b50-b230-4f94-bd39-7d66e4601e08\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.462090 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5nr9\" (UniqueName: \"kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9\") pod \"22907b50-b230-4f94-bd39-7d66e4601e08\" (UID: \"22907b50-b230-4f94-bd39-7d66e4601e08\") " Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.471183 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9" (OuterVolumeSpecName: "kube-api-access-r5nr9") pod "22907b50-b230-4f94-bd39-7d66e4601e08" (UID: "22907b50-b230-4f94-bd39-7d66e4601e08"). InnerVolumeSpecName "kube-api-access-r5nr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.506158 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81" (OuterVolumeSpecName: "mariadb-data") pod "22907b50-b230-4f94-bd39-7d66e4601e08" (UID: "22907b50-b230-4f94-bd39-7d66e4601e08"). InnerVolumeSpecName "pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.568312 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5nr9\" (UniqueName: \"kubernetes.io/projected/22907b50-b230-4f94-bd39-7d66e4601e08-kube-api-access-r5nr9\") on node \"crc\" DevicePath \"\"" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.568414 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") on node \"crc\" " Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.618188 4837 generic.go:334] "Generic (PLEG): container finished" podID="22907b50-b230-4f94-bd39-7d66e4601e08" containerID="ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86" exitCode=137 Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.618254 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"22907b50-b230-4f94-bd39-7d66e4601e08","Type":"ContainerDied","Data":"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86"} Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.618292 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"22907b50-b230-4f94-bd39-7d66e4601e08","Type":"ContainerDied","Data":"516954e5bb8b2779be8ed5a3a445596ea05f51ecf5164beab550a5ec221225a9"} Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.618308 4837 scope.go:117] "RemoveContainer" containerID="ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.618888 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.629527 4837 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.629801 4837 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81") on node "crc" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.667103 4837 scope.go:117] "RemoveContainer" containerID="ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86" Oct 01 10:03:50 crc kubenswrapper[4837]: E1001 10:03:50.668299 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86\": container with ID starting with ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86 not found: ID does not exist" containerID="ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.668353 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86"} err="failed to get container status \"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86\": rpc error: code = NotFound desc = could not find container \"ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86\": container with ID starting with ae00b649469d97214e197380d1de64766ac82557776a2ec222e2ab94aee50c86 not found: ID does not exist" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.670277 4837 reconciler_common.go:293] "Volume detached for volume \"pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f3bdaf7-dca8-49a1-89da-7ad2353d5d81\") on node \"crc\" DevicePath \"\"" Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.699100 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 10:03:50 crc kubenswrapper[4837]: I1001 10:03:50.714392 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 01 10:03:51 crc kubenswrapper[4837]: I1001 10:03:51.370729 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 01 10:03:51 crc kubenswrapper[4837]: I1001 10:03:51.371224 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="8e0d4988-7106-406d-aaf1-02b307230818" containerName="adoption" containerID="cri-o://2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861" gracePeriod=30 Oct 01 10:03:51 crc kubenswrapper[4837]: I1001 10:03:51.837954 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22907b50-b230-4f94-bd39-7d66e4601e08" path="/var/lib/kubelet/pods/22907b50-b230-4f94-bd39-7d66e4601e08/volumes" Oct 01 10:04:01 crc kubenswrapper[4837]: I1001 10:04:01.817010 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:04:01 crc kubenswrapper[4837]: E1001 10:04:01.818100 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:04:14 crc kubenswrapper[4837]: I1001 10:04:14.817338 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:04:14 crc kubenswrapper[4837]: E1001 10:04:14.818139 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.915380 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.987254 4837 generic.go:334] "Generic (PLEG): container finished" podID="8e0d4988-7106-406d-aaf1-02b307230818" containerID="2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861" exitCode=137 Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.987303 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"8e0d4988-7106-406d-aaf1-02b307230818","Type":"ContainerDied","Data":"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861"} Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.987331 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"8e0d4988-7106-406d-aaf1-02b307230818","Type":"ContainerDied","Data":"675ba22379501f672d50919a77e1b503f0fdfb769db3a3e45b8ba4f4c4372bd7"} Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.987350 4837 scope.go:117] "RemoveContainer" containerID="2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861" Oct 01 10:04:21 crc kubenswrapper[4837]: I1001 10:04:21.987350 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.019575 4837 scope.go:117] "RemoveContainer" containerID="2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861" Oct 01 10:04:22 crc kubenswrapper[4837]: E1001 10:04:22.020261 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861\": container with ID starting with 2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861 not found: ID does not exist" containerID="2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.020313 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861"} err="failed to get container status \"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861\": rpc error: code = NotFound desc = could not find container \"2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861\": container with ID starting with 2f9805b051de0cb7abe11165162a59eecce26f0e693c84de8b61c0eb8208a861 not found: ID does not exist" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.098168 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") pod \"8e0d4988-7106-406d-aaf1-02b307230818\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.098352 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert\") pod \"8e0d4988-7106-406d-aaf1-02b307230818\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.098511 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9hgf\" (UniqueName: \"kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf\") pod \"8e0d4988-7106-406d-aaf1-02b307230818\" (UID: \"8e0d4988-7106-406d-aaf1-02b307230818\") " Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.105905 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "8e0d4988-7106-406d-aaf1-02b307230818" (UID: "8e0d4988-7106-406d-aaf1-02b307230818"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.106835 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf" (OuterVolumeSpecName: "kube-api-access-m9hgf") pod "8e0d4988-7106-406d-aaf1-02b307230818" (UID: "8e0d4988-7106-406d-aaf1-02b307230818"). InnerVolumeSpecName "kube-api-access-m9hgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.123648 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e" (OuterVolumeSpecName: "ovn-data") pod "8e0d4988-7106-406d-aaf1-02b307230818" (UID: "8e0d4988-7106-406d-aaf1-02b307230818"). InnerVolumeSpecName "pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.201127 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9hgf\" (UniqueName: \"kubernetes.io/projected/8e0d4988-7106-406d-aaf1-02b307230818-kube-api-access-m9hgf\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.201213 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") on node \"crc\" " Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.201241 4837 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/8e0d4988-7106-406d-aaf1-02b307230818-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.229615 4837 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.229902 4837 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e") on node "crc" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.303889 4837 reconciler_common.go:293] "Volume detached for volume \"pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1f1c9c3-d0e9-409d-86d9-89d3d509458e\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.325179 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 01 10:04:22 crc kubenswrapper[4837]: I1001 10:04:22.336284 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 01 10:04:23 crc kubenswrapper[4837]: I1001 10:04:23.834516 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0d4988-7106-406d-aaf1-02b307230818" path="/var/lib/kubelet/pods/8e0d4988-7106-406d-aaf1-02b307230818/volumes" Oct 01 10:04:25 crc kubenswrapper[4837]: I1001 10:04:25.816507 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:04:25 crc kubenswrapper[4837]: E1001 10:04:25.817428 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.908967 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-9xxdk"] Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910059 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0d4988-7106-406d-aaf1-02b307230818" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910080 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0d4988-7106-406d-aaf1-02b307230818" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910113 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="extract-utilities" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910122 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="extract-utilities" Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910162 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22907b50-b230-4f94-bd39-7d66e4601e08" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910170 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="22907b50-b230-4f94-bd39-7d66e4601e08" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910187 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="extract-content" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910195 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="extract-content" Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910214 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956da144-c603-4569-b09e-0f4bce2d6c15" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910224 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="956da144-c603-4569-b09e-0f4bce2d6c15" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 01 10:04:33 crc kubenswrapper[4837]: E1001 10:04:33.910236 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="registry-server" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910244 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="registry-server" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910471 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0d4988-7106-406d-aaf1-02b307230818" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910502 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd8be8e-3eb3-4a8a-9b5d-60c44a9e98d7" containerName="registry-server" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910524 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="22907b50-b230-4f94-bd39-7d66e4601e08" containerName="adoption" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.910543 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="956da144-c603-4569-b09e-0f4bce2d6c15" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.912329 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.916876 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.918344 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:04:33 crc kubenswrapper[4837]: I1001 10:04:33.922287 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-9xxdk"] Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.072093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.072561 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.072941 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjlb\" (UniqueName: \"kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.073206 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.073464 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.073908 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.074056 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175584 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175767 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjlb\" (UniqueName: \"kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175792 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175829 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175901 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.175945 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.176009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.178533 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.178622 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.187687 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.188496 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.188766 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.193975 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjlb\" (UniqueName: \"kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb\") pod \"swift-ring-rebalance-debug-9xxdk\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.245277 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:34 crc kubenswrapper[4837]: I1001 10:04:34.754857 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-9xxdk"] Oct 01 10:04:35 crc kubenswrapper[4837]: I1001 10:04:35.155400 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9xxdk" event={"ID":"6b8056d5-4212-445d-a359-cc0f32f308d9","Type":"ContainerStarted","Data":"b86b179c61ce08c1c682bdcbeb1fc44dc9d577a257d785b700d50bcb04e2e49f"} Oct 01 10:04:35 crc kubenswrapper[4837]: I1001 10:04:35.155770 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9xxdk" event={"ID":"6b8056d5-4212-445d-a359-cc0f32f308d9","Type":"ContainerStarted","Data":"dad5b19f5d0171f443c04368beb057cf80fe34b39c4b5791d43deea0de41bb0a"} Oct 01 10:04:35 crc kubenswrapper[4837]: I1001 10:04:35.186577 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-9xxdk" podStartSLOduration=2.186555827 podStartE2EDuration="2.186555827s" podCreationTimestamp="2025-10-01 10:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:04:35.178016946 +0000 UTC m=+10732.019624421" watchObservedRunningTime="2025-10-01 10:04:35.186555827 +0000 UTC m=+10732.028163292" Oct 01 10:04:36 crc kubenswrapper[4837]: I1001 10:04:36.171084 4837 generic.go:334] "Generic (PLEG): container finished" podID="6b8056d5-4212-445d-a359-cc0f32f308d9" containerID="b86b179c61ce08c1c682bdcbeb1fc44dc9d577a257d785b700d50bcb04e2e49f" exitCode=0 Oct 01 10:04:36 crc kubenswrapper[4837]: I1001 10:04:36.171192 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9xxdk" event={"ID":"6b8056d5-4212-445d-a359-cc0f32f308d9","Type":"ContainerDied","Data":"b86b179c61ce08c1c682bdcbeb1fc44dc9d577a257d785b700d50bcb04e2e49f"} Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.607427 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.662050 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-9xxdk"] Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.680116 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-9xxdk"] Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761571 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761611 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761664 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkjlb\" (UniqueName: \"kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761775 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761824 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761886 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.761954 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift\") pod \"6b8056d5-4212-445d-a359-cc0f32f308d9\" (UID: \"6b8056d5-4212-445d-a359-cc0f32f308d9\") " Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.763000 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.763687 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.768465 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb" (OuterVolumeSpecName: "kube-api-access-qkjlb") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "kube-api-access-qkjlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.791304 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts" (OuterVolumeSpecName: "scripts") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.793979 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.807885 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.809473 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6b8056d5-4212-445d-a359-cc0f32f308d9" (UID: "6b8056d5-4212-445d-a359-cc0f32f308d9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.834955 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b8056d5-4212-445d-a359-cc0f32f308d9" path="/var/lib/kubelet/pods/6b8056d5-4212-445d-a359-cc0f32f308d9/volumes" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868246 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868331 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6b8056d5-4212-445d-a359-cc0f32f308d9-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868352 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868370 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868392 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkjlb\" (UniqueName: \"kubernetes.io/projected/6b8056d5-4212-445d-a359-cc0f32f308d9-kube-api-access-qkjlb\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868411 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6b8056d5-4212-445d-a359-cc0f32f308d9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:37 crc kubenswrapper[4837]: I1001 10:04:37.868430 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8056d5-4212-445d-a359-cc0f32f308d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:38 crc kubenswrapper[4837]: I1001 10:04:38.200862 4837 scope.go:117] "RemoveContainer" containerID="b86b179c61ce08c1c682bdcbeb1fc44dc9d577a257d785b700d50bcb04e2e49f" Oct 01 10:04:38 crc kubenswrapper[4837]: I1001 10:04:38.201097 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9xxdk" Oct 01 10:04:38 crc kubenswrapper[4837]: I1001 10:04:38.816515 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:04:38 crc kubenswrapper[4837]: E1001 10:04:38.817104 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.058311 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-28njj"] Oct 01 10:04:39 crc kubenswrapper[4837]: E1001 10:04:39.059202 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b8056d5-4212-445d-a359-cc0f32f308d9" containerName="swift-ring-rebalance" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.059289 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b8056d5-4212-445d-a359-cc0f32f308d9" containerName="swift-ring-rebalance" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.059643 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b8056d5-4212-445d-a359-cc0f32f308d9" containerName="swift-ring-rebalance" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.060590 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.064818 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.064823 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.073626 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-28njj"] Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099232 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftpl7\" (UniqueName: \"kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099296 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099357 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099412 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099497 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.099617 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.201925 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.201984 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202023 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202082 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftpl7\" (UniqueName: \"kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202106 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202149 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202186 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.202973 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.203331 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.203395 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.205969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.207005 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.208103 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.224178 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftpl7\" (UniqueName: \"kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7\") pod \"swift-ring-rebalance-debug-28njj\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.402886 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:04:39 crc kubenswrapper[4837]: I1001 10:04:39.959449 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-28njj"] Oct 01 10:04:40 crc kubenswrapper[4837]: I1001 10:04:40.234032 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-28njj" event={"ID":"36009b4a-6745-4fde-acd1-a113fa89a988","Type":"ContainerStarted","Data":"6afe04075f408846bdea9fa739447a8944873f6952e85c534038b6d3c5196b37"} Oct 01 10:04:41 crc kubenswrapper[4837]: I1001 10:04:41.246938 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-28njj" event={"ID":"36009b4a-6745-4fde-acd1-a113fa89a988","Type":"ContainerStarted","Data":"c8cf2bf651fae65a77b6c46a6bd79ec83f94f398ebdf6c5c5ef707903621ce36"} Oct 01 10:04:41 crc kubenswrapper[4837]: I1001 10:04:41.271080 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-28njj" podStartSLOduration=2.271060414 podStartE2EDuration="2.271060414s" podCreationTimestamp="2025-10-01 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:04:41.265012703 +0000 UTC m=+10738.106620208" watchObservedRunningTime="2025-10-01 10:04:41.271060414 +0000 UTC m=+10738.112667869" Oct 01 10:04:49 crc kubenswrapper[4837]: I1001 10:04:49.819277 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:04:49 crc kubenswrapper[4837]: E1001 10:04:49.820267 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:05:02 crc kubenswrapper[4837]: I1001 10:05:02.816110 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:05:02 crc kubenswrapper[4837]: E1001 10:05:02.817208 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:05:06 crc kubenswrapper[4837]: I1001 10:05:06.567209 4837 generic.go:334] "Generic (PLEG): container finished" podID="36009b4a-6745-4fde-acd1-a113fa89a988" containerID="c8cf2bf651fae65a77b6c46a6bd79ec83f94f398ebdf6c5c5ef707903621ce36" exitCode=0 Oct 01 10:05:06 crc kubenswrapper[4837]: I1001 10:05:06.567414 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-28njj" event={"ID":"36009b4a-6745-4fde-acd1-a113fa89a988","Type":"ContainerDied","Data":"c8cf2bf651fae65a77b6c46a6bd79ec83f94f398ebdf6c5c5ef707903621ce36"} Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.344363 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.389598 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-28njj"] Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.398975 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-28njj"] Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.503601 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.503678 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftpl7\" (UniqueName: \"kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.503769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.503897 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.503987 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.504087 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.504116 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts\") pod \"36009b4a-6745-4fde-acd1-a113fa89a988\" (UID: \"36009b4a-6745-4fde-acd1-a113fa89a988\") " Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.506377 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.506630 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.512482 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7" (OuterVolumeSpecName: "kube-api-access-ftpl7") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "kube-api-access-ftpl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.539429 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.550385 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.554724 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.556479 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts" (OuterVolumeSpecName: "scripts") pod "36009b4a-6745-4fde-acd1-a113fa89a988" (UID: "36009b4a-6745-4fde-acd1-a113fa89a988"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.596344 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6afe04075f408846bdea9fa739447a8944873f6952e85c534038b6d3c5196b37" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.596421 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-28njj" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611598 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611634 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftpl7\" (UniqueName: \"kubernetes.io/projected/36009b4a-6745-4fde-acd1-a113fa89a988-kube-api-access-ftpl7\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611646 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611658 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611670 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/36009b4a-6745-4fde-acd1-a113fa89a988-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611680 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36009b4a-6745-4fde-acd1-a113fa89a988-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:08 crc kubenswrapper[4837]: I1001 10:05:08.611706 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36009b4a-6745-4fde-acd1-a113fa89a988-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:09 crc kubenswrapper[4837]: I1001 10:05:09.827639 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36009b4a-6745-4fde-acd1-a113fa89a988" path="/var/lib/kubelet/pods/36009b4a-6745-4fde-acd1-a113fa89a988/volumes" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.733549 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 01 10:05:10 crc kubenswrapper[4837]: E1001 10:05:10.734403 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36009b4a-6745-4fde-acd1-a113fa89a988" containerName="swift-ring-rebalance" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.734419 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="36009b4a-6745-4fde-acd1-a113fa89a988" containerName="swift-ring-rebalance" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.734653 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="36009b4a-6745-4fde-acd1-a113fa89a988" containerName="swift-ring-rebalance" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.741445 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.744502 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.764264 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.783468 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-2"] Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.791524 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.796477 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-1"] Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.812101 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.825114 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-2"] Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.835934 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-1"] Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868124 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6b2n\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-kube-api-access-r6b2n\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868220 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-etc-swift\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868275 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868312 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-lock\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868385 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-lock\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868437 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-cache\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868466 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-etc-swift\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868500 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z88rl\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-kube-api-access-z88rl\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868539 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.868579 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-cache\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970459 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-lock\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970544 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-etc-swift\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970655 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-lock\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970726 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-cache\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970828 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-lock\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970878 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-cache\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970945 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-etc-swift\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.970987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzmv5\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-kube-api-access-bzmv5\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971047 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z88rl\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-kube-api-access-z88rl\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971137 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971272 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-cache\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971355 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6b2n\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-kube-api-access-r6b2n\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971456 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971498 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-etc-swift\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.971607 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.975232 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-cache\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.976264 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ac458f22-6672-45bf-89f6-d19c35debb4f-lock\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.976386 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-lock\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.976751 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f5622e0a-790d-453a-bdcc-86e778a90b61-cache\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.979551 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.979659 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/57d557992192dd11408890b3fc4b778da1c1fd320496252e2a01e9c7e416b2c6/globalmount\"" pod="openstack/swift-storage-2" Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.979905 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 10:05:10 crc kubenswrapper[4837]: I1001 10:05:10.980021 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5be63df46bb629b05134e9fcbb9d1c1c4df41aad77d41a13fb3dd518f8933112/globalmount\"" pod="openstack/swift-storage-0" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.073575 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzmv5\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-kube-api-access-bzmv5\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.073706 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.073795 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-etc-swift\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.073845 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-cache\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.073875 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-lock\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.074306 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-lock\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.074485 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b4549e0-58e4-4bf6-af2b-6fbad629511c-cache\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.075855 4837 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.075887 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0182456e7e7141f4ace57a1074e77477a5c67ae60ade04c402d4269b0b7a7acd/globalmount\"" pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.451817 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z88rl\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-kube-api-access-z88rl\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.451944 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6b2n\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-kube-api-access-r6b2n\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.452776 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ac458f22-6672-45bf-89f6-d19c35debb4f-etc-swift\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.452869 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5622e0a-790d-453a-bdcc-86e778a90b61-etc-swift\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.453321 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzmv5\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-kube-api-access-bzmv5\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.454594 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b4549e0-58e4-4bf6-af2b-6fbad629511c-etc-swift\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.503545 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4c62a133-03fd-4e24-878b-6de7456b29eb\") pod \"swift-storage-2\" (UID: \"ac458f22-6672-45bf-89f6-d19c35debb4f\") " pod="openstack/swift-storage-2" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.510022 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ba8299-3ce5-4704-a3d7-dda501292114\") pod \"swift-storage-0\" (UID: \"f5622e0a-790d-453a-bdcc-86e778a90b61\") " pod="openstack/swift-storage-0" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.513510 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-65b27bf6-a846-4447-b2eb-1481c04419c8\") pod \"swift-storage-1\" (UID: \"7b4549e0-58e4-4bf6-af2b-6fbad629511c\") " pod="openstack/swift-storage-1" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.724607 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.747077 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-2" Oct 01 10:05:11 crc kubenswrapper[4837]: I1001 10:05:11.758301 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-1" Oct 01 10:05:12 crc kubenswrapper[4837]: I1001 10:05:12.404179 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 10:05:12 crc kubenswrapper[4837]: I1001 10:05:12.658293 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"2bc2681f113140cead13bb820d521a0938c604607b46417f9611673e4cd39502"} Oct 01 10:05:12 crc kubenswrapper[4837]: I1001 10:05:12.763266 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-2"] Oct 01 10:05:12 crc kubenswrapper[4837]: W1001 10:05:12.772314 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac458f22_6672_45bf_89f6_d19c35debb4f.slice/crio-23f688c49d3cb304f1b6cbe46427e1f6bbb3dfbd49dbfae1c52b005323b30ea0 WatchSource:0}: Error finding container 23f688c49d3cb304f1b6cbe46427e1f6bbb3dfbd49dbfae1c52b005323b30ea0: Status 404 returned error can't find the container with id 23f688c49d3cb304f1b6cbe46427e1f6bbb3dfbd49dbfae1c52b005323b30ea0 Oct 01 10:05:13 crc kubenswrapper[4837]: I1001 10:05:13.673581 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"23f688c49d3cb304f1b6cbe46427e1f6bbb3dfbd49dbfae1c52b005323b30ea0"} Oct 01 10:05:13 crc kubenswrapper[4837]: I1001 10:05:13.701383 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-1"] Oct 01 10:05:13 crc kubenswrapper[4837]: W1001 10:05:13.780583 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b4549e0_58e4_4bf6_af2b_6fbad629511c.slice/crio-709c1f8a60780f2a5b965c767e0d91d14eb5f1b8a56e90c2a06433c4b8573bc3 WatchSource:0}: Error finding container 709c1f8a60780f2a5b965c767e0d91d14eb5f1b8a56e90c2a06433c4b8573bc3: Status 404 returned error can't find the container with id 709c1f8a60780f2a5b965c767e0d91d14eb5f1b8a56e90c2a06433c4b8573bc3 Oct 01 10:05:13 crc kubenswrapper[4837]: I1001 10:05:13.836074 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:05:13 crc kubenswrapper[4837]: E1001 10:05:13.836296 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.693147 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-zdpwn"] Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.713322 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"319d5c61eddd510ba984c9cfd275a78af76bb99741b60535f08d6dec03847d01"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.713380 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"2e8b7646482f76df317169427837873bffe0823c6c5c8d3ae62127f63fe8bab0"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.718596 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-zdpwn"] Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.719015 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"fb9cf9b79e8aba518e6f07123f99cdba7774c87e805b6b08e67008dbf75debc7"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.719060 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"eace926e845fbd5bb68c29575a3d19b32a471b046939a53e7679ccfcccf8d05b"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.732992 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"23fbe29eeb9157c72ea61f66bea431c9a6a7353bc579ee4bb34c6b2ce9c94443"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.733036 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"bbd25418e4f8fa6e4715b759c197e53fa1009a653c9605d3a6a357c768ce2f9f"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.733046 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"709c1f8a60780f2a5b965c767e0d91d14eb5f1b8a56e90c2a06433c4b8573bc3"} Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.743901 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-45l5p"] Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.745277 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.756819 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.757847 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.765908 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-45l5p"] Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.873872 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874020 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874077 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874120 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28xk5\" (UniqueName: \"kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874188 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.874274 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976260 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976391 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976427 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976464 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976501 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28xk5\" (UniqueName: \"kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976546 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.976723 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.977360 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.977820 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.982254 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.982347 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.982752 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:14 crc kubenswrapper[4837]: I1001 10:05:14.993844 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28xk5\" (UniqueName: \"kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5\") pod \"swift-ring-rebalance-45l5p\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.144807 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.660795 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-45l5p"] Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.770002 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"eb94c5c3321a7690d603fb15100ead04333430c954a92b818e3db97c25d3c3c4"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.770043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"2345154f0a1053adf3c548552077520a54e98f7b0b7f1cdd49e3d21eaeb337cc"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.773509 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"126e1e9a3bcca373251d75336a640001b18c3967f956d71cf91906c94e31103c"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.773535 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"1b295faf96ef78c882f1d0a699676d847738ed60575509c71e40a35aac37b791"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.777049 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"62664ee52738ee0a97cea307592bc9fe6da8dce08f6303ce308dd141521185d6"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.777074 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"7dd526ab4580f25406945419013858090c6a27c05baa74c6789489efd8d4946b"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.778323 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-45l5p" event={"ID":"5fc49e27-d900-4566-a796-c1d6e72e2444","Type":"ContainerStarted","Data":"7fd49e18af2435a00d67134eadcacb75f7e8cfb9dea6f44662abbf8cf68cbb18"} Oct 01 10:05:15 crc kubenswrapper[4837]: I1001 10:05:15.836617 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e55925e1-26b2-42da-a0b6-dacb251e66ed" path="/var/lib/kubelet/pods/e55925e1-26b2-42da-a0b6-dacb251e66ed/volumes" Oct 01 10:05:16 crc kubenswrapper[4837]: I1001 10:05:16.798892 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-45l5p" event={"ID":"5fc49e27-d900-4566-a796-c1d6e72e2444","Type":"ContainerStarted","Data":"8925373551b2a87ab457345804c64d8feccbcbde1e60e63d58f203b27c4dc8c5"} Oct 01 10:05:16 crc kubenswrapper[4837]: I1001 10:05:16.827455 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-45l5p" podStartSLOduration=2.827436668 podStartE2EDuration="2.827436668s" podCreationTimestamp="2025-10-01 10:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:05:16.818410284 +0000 UTC m=+10773.660017739" watchObservedRunningTime="2025-10-01 10:05:16.827436668 +0000 UTC m=+10773.669044123" Oct 01 10:05:16 crc kubenswrapper[4837]: I1001 10:05:16.992802 4837 scope.go:117] "RemoveContainer" containerID="4b9a7ff4a5e29f33a867427289d69fb34ef1aca2e3a5c9a1a4c1cab4d46669b1" Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.839539 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"962fcc0db286c0a0a53075471ee2c8c2b729a6516060ff8010716c75a21ba1af"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.839894 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"ee496f44b564d962796a4c1adc14f0c5374b0bd7ca26598a4c89b19140c6aa36"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.854625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"c8727c76cb41eb9c414b658df2b2b95dc0830c6ce4d811dfd5f83e034e79ed4f"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.854669 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"35f8b33cc15261352695ee5cf69e111688a01bd53b231a6b0ed2a2de2befc167"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.854678 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"4ab2d2567e868e8a05b34973f116526b7e67e67688e1dc4d9bcb47c50059a7fd"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.857619 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"80a4d070fd4826a86c8662abffa91d1a70199de4118913a147ad6923493952a3"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.857708 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"a170cbf0b0f16e7aa73b907391afda5ed5bdbe308900d769ac798ea20641d879"} Oct 01 10:05:17 crc kubenswrapper[4837]: I1001 10:05:17.857721 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"4dab5642e584a59660343ebef83fc39c08e53816e7ebb939f3ad83fa70d5d096"} Oct 01 10:05:18 crc kubenswrapper[4837]: I1001 10:05:18.904067 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"ade23b7cb889102f955c069b074c03fb53fdbf80c540241bd6e6c484cf2ca2fd"} Oct 01 10:05:18 crc kubenswrapper[4837]: I1001 10:05:18.911097 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"645ae3348cc650e987ce9a31d949afb65c21d3fd5f7ecae762286c3942d2041c"} Oct 01 10:05:18 crc kubenswrapper[4837]: I1001 10:05:18.911163 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"c6ecc87e2b0f907910be184cbc947196464bfc2049636ea3aa8217c6366375cf"} Oct 01 10:05:18 crc kubenswrapper[4837]: I1001 10:05:18.916439 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"40c6f42194532220b9894ff19708f29e51b7d4d821622d9046bdf9ae7901a8d1"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.970628 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"25d4a11cb1de152b7dc81764c8c5ca31ecd64caee2e6e93b92b5bac0056834fc"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.971293 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"d9244cd1a1019cb84bc0fb4fce5ee629c5f691548cc318b79317a76b016d1118"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.971306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"73b71bee0028699b63098ed784af4ef8747971c42dfc8c5d1dbccabf884d3509"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.994618 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"db7c3230b443608a7297df3ec3e389cbd24a3591c52d530dc6e0c523e6da6cea"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.994667 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"08ff2a2d57a0361b39bc94751abbec9cb2a77c91f1f0554aa0f5d5863d35ceed"} Oct 01 10:05:19 crc kubenswrapper[4837]: I1001 10:05:19.994677 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"a77d23475c432c87d9a659fa96e3052329089ad13ca544247001bc882ad9c68d"} Oct 01 10:05:20 crc kubenswrapper[4837]: I1001 10:05:20.041120 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"fdac8763806b8f5159d003e7ba736f59bb87c28f2fbb3dca4ad4d00e93a36249"} Oct 01 10:05:20 crc kubenswrapper[4837]: I1001 10:05:20.041259 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"efe20991282f169a6d134a1d0a57d41b8f43c18cd008b0bfcbae986cbdf40870"} Oct 01 10:05:20 crc kubenswrapper[4837]: I1001 10:05:20.041323 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"c3ce00163356969bbc9ec9168f11d1800b4e4b14bd276e9fa842934963159256"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.060960 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"13e510c245e5da80533f0a53397df80687f339a9d9efb475daa0e3c705f37ec5"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.061255 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"85107b56b8428a8c447890580f70138a7c283c0a49feccfa698a02aa20a10b7c"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.061265 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"0ca95a79dc095d003ad96cae828f0fc4bddc5686f8bd20b5d79db66602ca1f9b"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.067655 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"8d5cbd9111f7c9c169fc79acdb92d306f878c861b18166720146ca61f9d5ce2f"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.067752 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"72d8cc076f1c319bfa7627ed7ca3ea3d665de8bb46d6617a5f6ae4b855f3d9dc"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.067770 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"000ab15465461110aaff591d9e9230e9b997aa5c30aa5194d0725df358713693"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.073118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"c05f47ade19a3b85412006e13cb1ebfaa7a639f7d3cefe8791c1f8de83678013"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.073144 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"d83c785dfd74141113716d52427f135fda586205164e71a6d98fe0d4f7625e8c"} Oct 01 10:05:21 crc kubenswrapper[4837]: I1001 10:05:21.073154 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"57bd5316f42c4c734b494315332c70f15e4f8588bf90200bec81cdcb794c733e"} Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.089600 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"ac458f22-6672-45bf-89f6-d19c35debb4f","Type":"ContainerStarted","Data":"2878bdcd15fdba8bafc7e18c8c17c45f6329b56bd26e018d3743df61826ab3fe"} Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.102385 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f5622e0a-790d-453a-bdcc-86e778a90b61","Type":"ContainerStarted","Data":"ba37a66a43b1df442d468dd24649ba3b643160ed6fb63e939a155c7be2b63b01"} Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.113772 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"7b4549e0-58e4-4bf6-af2b-6fbad629511c","Type":"ContainerStarted","Data":"0023fd7652d8df0621fb06c4d78dc30eed388fc2126e764d96b19e4f85a695a2"} Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.159189 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-2" podStartSLOduration=7.104932591 podStartE2EDuration="13.159167508s" podCreationTimestamp="2025-10-01 10:05:09 +0000 UTC" firstStartedPulling="2025-10-01 10:05:12.776700515 +0000 UTC m=+10769.618307970" lastFinishedPulling="2025-10-01 10:05:18.830935432 +0000 UTC m=+10775.672542887" observedRunningTime="2025-10-01 10:05:22.15120338 +0000 UTC m=+10778.992810835" watchObservedRunningTime="2025-10-01 10:05:22.159167508 +0000 UTC m=+10779.000774963" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.207923 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-1" podStartSLOduration=8.141994001 podStartE2EDuration="13.207896593s" podCreationTimestamp="2025-10-01 10:05:09 +0000 UTC" firstStartedPulling="2025-10-01 10:05:13.782816671 +0000 UTC m=+10770.624424126" lastFinishedPulling="2025-10-01 10:05:18.848719263 +0000 UTC m=+10775.690326718" observedRunningTime="2025-10-01 10:05:22.197097576 +0000 UTC m=+10779.038705031" watchObservedRunningTime="2025-10-01 10:05:22.207896593 +0000 UTC m=+10779.049504048" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.256453 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=6.795964305 podStartE2EDuration="13.256435124s" podCreationTimestamp="2025-10-01 10:05:09 +0000 UTC" firstStartedPulling="2025-10-01 10:05:12.395315218 +0000 UTC m=+10769.236922713" lastFinishedPulling="2025-10-01 10:05:18.855786077 +0000 UTC m=+10775.697393532" observedRunningTime="2025-10-01 10:05:22.244641603 +0000 UTC m=+10779.086249058" watchObservedRunningTime="2025-10-01 10:05:22.256435124 +0000 UTC m=+10779.098042569" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.564716 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6986dc8477-hpjcw"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.567481 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.572425 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.618425 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6986dc8477-hpjcw"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.654973 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6986dc8477-hpjcw"] Oct 01 10:05:22 crc kubenswrapper[4837]: E1001 10:05:22.657140 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-9dp77 openstack-cell1 openstack-networker ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[dns-svc dns-swift-storage-0 kube-api-access-9dp77 openstack-cell1 openstack-networker ovsdbserver-nb ovsdbserver-sb]: context canceled" pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" podUID="5cc1281e-e04a-49f6-a44e-8f3466da6c53" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.673961 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dp77\" (UniqueName: \"kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674013 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674050 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674079 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674094 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674115 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674203 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.674238 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.679389 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64c4696475-qx7h5"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.699602 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.704137 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-1" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.704220 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-2" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.708244 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c4696475-qx7h5"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.775789 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-2\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.775896 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dp77\" (UniqueName: \"kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.775941 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.775967 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.775999 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjl5k\" (UniqueName: \"kubernetes.io/projected/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-kube-api-access-wjl5k\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776033 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776066 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776086 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776108 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-0\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776124 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776165 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776230 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-config\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776264 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776294 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-networker\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776358 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776428 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776462 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-cell1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.776504 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-svc\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.777109 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.777149 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.777381 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.778106 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.778179 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.778321 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.780741 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.812598 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dp77\" (UniqueName: \"kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77\") pod \"dnsmasq-dns-6986dc8477-hpjcw\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.878910 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.878981 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-cell1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879053 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-svc\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879131 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-2\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879250 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879309 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjl5k\" (UniqueName: \"kubernetes.io/projected/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-kube-api-access-wjl5k\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879421 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-0\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879502 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879584 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-config\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.879676 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-networker\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.881009 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-networker\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.882291 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.883011 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-openstack-cell1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.884921 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.885195 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-2\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.885357 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-svc\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.885404 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-config\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.885645 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-1\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.886349 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-dns-swift-storage-0\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.894737 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.897516 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.918017 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.929548 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjl5k\" (UniqueName: \"kubernetes.io/projected/dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546-kube-api-access-wjl5k\") pod \"dnsmasq-dns-64c4696475-qx7h5\" (UID: \"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546\") " pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.981791 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982015 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfxvh\" (UniqueName: \"kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982124 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982221 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982244 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982294 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:22 crc kubenswrapper[4837]: I1001 10:05:22.982621 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.032496 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.085900 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086500 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfxvh\" (UniqueName: \"kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086566 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086615 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086726 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.086820 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.088623 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.088878 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.089263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.090854 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.090977 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.104135 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.106107 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfxvh\" (UniqueName: \"kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh\") pod \"swift-ring-rebalance-debug-gd4vx\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.134005 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.234512 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.236543 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.306414 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.314908 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.314941 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.314961 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.315121 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dp77\" (UniqueName: \"kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.315204 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.315242 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.315286 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc\") pod \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\" (UID: \"5cc1281e-e04a-49f6-a44e-8f3466da6c53\") " Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.307150 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.316106 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.316558 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.316940 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config" (OuterVolumeSpecName: "config") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.317251 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.317391 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.317462 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.321441 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77" (OuterVolumeSpecName: "kube-api-access-9dp77") pod "5cc1281e-e04a-49f6-a44e-8f3466da6c53" (UID: "5cc1281e-e04a-49f6-a44e-8f3466da6c53"). InnerVolumeSpecName "kube-api-access-9dp77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419556 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-config\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419635 4837 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419653 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dp77\" (UniqueName: \"kubernetes.io/projected/5cc1281e-e04a-49f6-a44e-8f3466da6c53-kube-api-access-9dp77\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419668 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419683 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419723 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419739 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.419755 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cc1281e-e04a-49f6-a44e-8f3466da6c53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.594897 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c4696475-qx7h5"] Oct 01 10:05:23 crc kubenswrapper[4837]: W1001 10:05:23.604216 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc42f6d7_dda3_4968_a4fc_d9b5ec2d5546.slice/crio-1c8e7f64d78520a599d0f0b3d245cd52f76cd8eb0bb80f35323a6393703291f0 WatchSource:0}: Error finding container 1c8e7f64d78520a599d0f0b3d245cd52f76cd8eb0bb80f35323a6393703291f0: Status 404 returned error can't find the container with id 1c8e7f64d78520a599d0f0b3d245cd52f76cd8eb0bb80f35323a6393703291f0 Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.755914 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:23 crc kubenswrapper[4837]: W1001 10:05:23.757168 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0462651_fcdf_4d08_bcfe_b82e0ccc9c5e.slice/crio-5bfe5eb08a70ea23c1c12eabf8ba3ea62de79091988f6804e7c9e6ae88b4d381 WatchSource:0}: Error finding container 5bfe5eb08a70ea23c1c12eabf8ba3ea62de79091988f6804e7c9e6ae88b4d381: Status 404 returned error can't find the container with id 5bfe5eb08a70ea23c1c12eabf8ba3ea62de79091988f6804e7c9e6ae88b4d381 Oct 01 10:05:23 crc kubenswrapper[4837]: I1001 10:05:23.930724 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.150533 4837 generic.go:334] "Generic (PLEG): container finished" podID="dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546" containerID="055419310a083e7953eb6437922cb3f8a37b9c47e89fa669c4cfe61cc7873aeb" exitCode=0 Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.150833 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" event={"ID":"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546","Type":"ContainerDied","Data":"055419310a083e7953eb6437922cb3f8a37b9c47e89fa669c4cfe61cc7873aeb"} Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.150903 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" event={"ID":"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546","Type":"ContainerStarted","Data":"1c8e7f64d78520a599d0f0b3d245cd52f76cd8eb0bb80f35323a6393703291f0"} Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.159033 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6986dc8477-hpjcw" Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.160430 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-gd4vx" event={"ID":"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e","Type":"ContainerStarted","Data":"5bfe5eb08a70ea23c1c12eabf8ba3ea62de79091988f6804e7c9e6ae88b4d381"} Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.402652 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6986dc8477-hpjcw"] Oct 01 10:05:24 crc kubenswrapper[4837]: I1001 10:05:24.450883 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6986dc8477-hpjcw"] Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.173147 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-gd4vx" event={"ID":"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e","Type":"ContainerDied","Data":"3166569d1d5251b10b0d0b927c5b3cfb7a066bacc861d5cad529ff1d3438e001"} Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.172979 4837 generic.go:334] "Generic (PLEG): container finished" podID="f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" containerID="3166569d1d5251b10b0d0b927c5b3cfb7a066bacc861d5cad529ff1d3438e001" exitCode=2 Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.175521 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" event={"ID":"dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546","Type":"ContainerStarted","Data":"05b0ed43547e9164cc838bfa7644f09f73e505cb038ce0e79b6e14cb63983233"} Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.177067 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.242367 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.257622 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" podStartSLOduration=3.255854662 podStartE2EDuration="3.255854662s" podCreationTimestamp="2025-10-01 10:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:05:25.224518927 +0000 UTC m=+10782.066126392" watchObservedRunningTime="2025-10-01 10:05:25.255854662 +0000 UTC m=+10782.097462117" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.278426 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-gd4vx"] Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.833497 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc1281e-e04a-49f6-a44e-8f3466da6c53" path="/var/lib/kubelet/pods/5cc1281e-e04a-49f6-a44e-8f3466da6c53/volumes" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.840122 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.978644 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.978769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.978889 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.978969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.979059 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.979122 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.979172 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfxvh\" (UniqueName: \"kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh\") pod \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\" (UID: \"f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e\") " Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.979367 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.980223 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.980774 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:25 crc kubenswrapper[4837]: I1001 10:05:25.985951 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh" (OuterVolumeSpecName: "kube-api-access-lfxvh") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "kube-api-access-lfxvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.019828 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.024260 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.052582 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.055122 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts" (OuterVolumeSpecName: "scripts") pod "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" (UID: "f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082039 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082305 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082366 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082423 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082483 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.082536 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfxvh\" (UniqueName: \"kubernetes.io/projected/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e-kube-api-access-lfxvh\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.194200 4837 scope.go:117] "RemoveContainer" containerID="3166569d1d5251b10b0d0b927c5b3cfb7a066bacc861d5cad529ff1d3438e001" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.194217 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gd4vx" Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.197329 4837 generic.go:334] "Generic (PLEG): container finished" podID="5fc49e27-d900-4566-a796-c1d6e72e2444" containerID="8925373551b2a87ab457345804c64d8feccbcbde1e60e63d58f203b27c4dc8c5" exitCode=0 Oct 01 10:05:26 crc kubenswrapper[4837]: I1001 10:05:26.198372 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-45l5p" event={"ID":"5fc49e27-d900-4566-a796-c1d6e72e2444","Type":"ContainerDied","Data":"8925373551b2a87ab457345804c64d8feccbcbde1e60e63d58f203b27c4dc8c5"} Oct 01 10:05:27 crc kubenswrapper[4837]: I1001 10:05:27.816254 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:05:27 crc kubenswrapper[4837]: I1001 10:05:27.834028 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" path="/var/lib/kubelet/pods/f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e/volumes" Oct 01 10:05:27 crc kubenswrapper[4837]: I1001 10:05:27.873287 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.027769 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.027856 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.027969 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28xk5\" (UniqueName: \"kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.028306 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.028887 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.028956 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.029092 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.029138 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts\") pod \"5fc49e27-d900-4566-a796-c1d6e72e2444\" (UID: \"5fc49e27-d900-4566-a796-c1d6e72e2444\") " Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.029819 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.029881 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.033515 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5" (OuterVolumeSpecName: "kube-api-access-28xk5") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "kube-api-access-28xk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.077060 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts" (OuterVolumeSpecName: "scripts") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.078210 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.082811 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.091727 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5fc49e27-d900-4566-a796-c1d6e72e2444" (UID: "5fc49e27-d900-4566-a796-c1d6e72e2444"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131439 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131473 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fc49e27-d900-4566-a796-c1d6e72e2444-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131483 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131492 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28xk5\" (UniqueName: \"kubernetes.io/projected/5fc49e27-d900-4566-a796-c1d6e72e2444-kube-api-access-28xk5\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131501 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fc49e27-d900-4566-a796-c1d6e72e2444-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.131508 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fc49e27-d900-4566-a796-c1d6e72e2444-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.258350 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b"} Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.261591 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-45l5p" event={"ID":"5fc49e27-d900-4566-a796-c1d6e72e2444","Type":"ContainerDied","Data":"7fd49e18af2435a00d67134eadcacb75f7e8cfb9dea6f44662abbf8cf68cbb18"} Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.261618 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fd49e18af2435a00d67134eadcacb75f7e8cfb9dea6f44662abbf8cf68cbb18" Oct 01 10:05:28 crc kubenswrapper[4837]: I1001 10:05:28.261731 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-45l5p" Oct 01 10:05:28 crc kubenswrapper[4837]: E1001 10:05:28.371406 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc49e27_d900_4566_a796_c1d6e72e2444.slice/crio-7fd49e18af2435a00d67134eadcacb75f7e8cfb9dea6f44662abbf8cf68cbb18\": RecentStats: unable to find data in memory cache]" Oct 01 10:05:33 crc kubenswrapper[4837]: I1001 10:05:33.036049 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64c4696475-qx7h5" Oct 01 10:05:33 crc kubenswrapper[4837]: I1001 10:05:33.126162 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 10:05:33 crc kubenswrapper[4837]: I1001 10:05:33.126593 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="dnsmasq-dns" containerID="cri-o://7587a55738e24cf1357a086d0af7f30cf8e438a689be538577d1f89ebd3e96e1" gracePeriod=10 Oct 01 10:05:33 crc kubenswrapper[4837]: I1001 10:05:33.387047 4837 generic.go:334] "Generic (PLEG): container finished" podID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerID="7587a55738e24cf1357a086d0af7f30cf8e438a689be538577d1f89ebd3e96e1" exitCode=0 Oct 01 10:05:33 crc kubenswrapper[4837]: I1001 10:05:33.387387 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" event={"ID":"2f6ede11-1c94-4644-8a8d-b508c1d48cca","Type":"ContainerDied","Data":"7587a55738e24cf1357a086d0af7f30cf8e438a689be538577d1f89ebd3e96e1"} Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.034175 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183366 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183460 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183520 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183571 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183756 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183817 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnw2p\" (UniqueName: \"kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.183905 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker\") pod \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\" (UID: \"2f6ede11-1c94-4644-8a8d-b508c1d48cca\") " Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.196388 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p" (OuterVolumeSpecName: "kube-api-access-lnw2p") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "kube-api-access-lnw2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.295600 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnw2p\" (UniqueName: \"kubernetes.io/projected/2f6ede11-1c94-4644-8a8d-b508c1d48cca-kube-api-access-lnw2p\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.296407 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.299004 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.308462 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.332727 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config" (OuterVolumeSpecName: "config") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.333556 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.341034 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2f6ede11-1c94-4644-8a8d-b508c1d48cca" (UID: "2f6ede11-1c94-4644-8a8d-b508c1d48cca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398518 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398552 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398566 4837 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-config\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398577 4837 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398588 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.398599 4837 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2f6ede11-1c94-4644-8a8d-b508c1d48cca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.423131 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" event={"ID":"2f6ede11-1c94-4644-8a8d-b508c1d48cca","Type":"ContainerDied","Data":"e6cf9d1fe3404cd95aba6aa5ab1f36d30fe23ee79a96ee2645bc83fc9ae04cd8"} Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.423195 4837 scope.go:117] "RemoveContainer" containerID="7587a55738e24cf1357a086d0af7f30cf8e438a689be538577d1f89ebd3e96e1" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.423405 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8594d5ff-7nl6w" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.458196 4837 scope.go:117] "RemoveContainer" containerID="df08dd925c5831cb249c3a3bd952809b010c8c747556c264b64ae57c5481ca6d" Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.469156 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 10:05:34 crc kubenswrapper[4837]: I1001 10:05:34.480486 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d8594d5ff-7nl6w"] Oct 01 10:05:35 crc kubenswrapper[4837]: I1001 10:05:35.829667 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" path="/var/lib/kubelet/pods/2f6ede11-1c94-4644-8a8d-b508c1d48cca/volumes" Oct 01 10:06:17 crc kubenswrapper[4837]: I1001 10:06:17.209947 4837 scope.go:117] "RemoveContainer" containerID="4043cc2583fbbfaa473398c163a80f2847d49053ae389a876b10fcf78adb0bb5" Oct 01 10:06:17 crc kubenswrapper[4837]: I1001 10:06:17.238110 4837 scope.go:117] "RemoveContainer" containerID="85d7121d3a012da5c6b778ea4ca2ff2d54761bed5515a9b438ac0961f05b553e" Oct 01 10:06:17 crc kubenswrapper[4837]: I1001 10:06:17.264289 4837 scope.go:117] "RemoveContainer" containerID="62b720018f4e924dc541a74e11d5644c14aa5ce9bb72a58de710f9998a605d6a" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.435488 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-2znrg"] Oct 01 10:06:25 crc kubenswrapper[4837]: E1001 10:06:25.436519 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="init" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436533 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="init" Oct 01 10:06:25 crc kubenswrapper[4837]: E1001 10:06:25.436562 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436572 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: E1001 10:06:25.436606 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc49e27-d900-4566-a796-c1d6e72e2444" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436614 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc49e27-d900-4566-a796-c1d6e72e2444" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: E1001 10:06:25.436637 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="dnsmasq-dns" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436647 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="dnsmasq-dns" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436893 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc49e27-d900-4566-a796-c1d6e72e2444" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436936 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0462651-fcdf-4d08-bcfe-b82e0ccc9c5e" containerName="swift-ring-rebalance" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.436963 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6ede11-1c94-4644-8a8d-b508c1d48cca" containerName="dnsmasq-dns" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.437834 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2znrg"] Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.437922 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.469318 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.469466 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.600327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.600683 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.600822 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.600986 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.601102 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.601181 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.601310 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2s8n\" (UniqueName: \"kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703061 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703145 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703238 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703277 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703307 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.703381 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2s8n\" (UniqueName: \"kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.704235 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.704330 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.704372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.704849 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.709048 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.715478 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.722209 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.724676 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2s8n\" (UniqueName: \"kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n\") pod \"swift-ring-rebalance-debug-2znrg\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:25 crc kubenswrapper[4837]: I1001 10:06:25.791949 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:26 crc kubenswrapper[4837]: I1001 10:06:26.255277 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2znrg"] Oct 01 10:06:26 crc kubenswrapper[4837]: W1001 10:06:26.262296 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89d7c14e_9ab5_45ec_8cae_6734847f9fa2.slice/crio-0665f7260cf617b49ada56fb50a9aed50a6ac95b0b3bfe1a5ea4fbeaabac88e2 WatchSource:0}: Error finding container 0665f7260cf617b49ada56fb50a9aed50a6ac95b0b3bfe1a5ea4fbeaabac88e2: Status 404 returned error can't find the container with id 0665f7260cf617b49ada56fb50a9aed50a6ac95b0b3bfe1a5ea4fbeaabac88e2 Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.075473 4837 generic.go:334] "Generic (PLEG): container finished" podID="89d7c14e-9ab5-45ec-8cae-6734847f9fa2" containerID="0bea40c3afc1c94782f1258dd226d84ae76dec151ed7e53f0c6730298a09d1ab" exitCode=0 Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.075511 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2znrg" event={"ID":"89d7c14e-9ab5-45ec-8cae-6734847f9fa2","Type":"ContainerDied","Data":"0bea40c3afc1c94782f1258dd226d84ae76dec151ed7e53f0c6730298a09d1ab"} Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.075845 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2znrg" event={"ID":"89d7c14e-9ab5-45ec-8cae-6734847f9fa2","Type":"ContainerStarted","Data":"0665f7260cf617b49ada56fb50a9aed50a6ac95b0b3bfe1a5ea4fbeaabac88e2"} Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.166052 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-2znrg"] Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.181355 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-2znrg"] Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.593598 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-sn8dr"] Oct 01 10:06:27 crc kubenswrapper[4837]: E1001 10:06:27.594248 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d7c14e-9ab5-45ec-8cae-6734847f9fa2" containerName="swift-ring-rebalance" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.594275 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d7c14e-9ab5-45ec-8cae-6734847f9fa2" containerName="swift-ring-rebalance" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.594611 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d7c14e-9ab5-45ec-8cae-6734847f9fa2" containerName="swift-ring-rebalance" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.595623 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.612396 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sn8dr"] Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.682763 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.682888 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wtlv\" (UniqueName: \"kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.683039 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.683078 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.683116 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.683206 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.683252 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785304 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785460 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785572 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785609 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wtlv\" (UniqueName: \"kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785737 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785772 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.785803 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.786642 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.787744 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.788102 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.791825 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.792070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.793330 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.803583 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wtlv\" (UniqueName: \"kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv\") pod \"swift-ring-rebalance-debug-sn8dr\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:27 crc kubenswrapper[4837]: I1001 10:06:27.954615 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.441769 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sn8dr"] Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.608718 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.803443 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2s8n\" (UniqueName: \"kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.803744 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.803898 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804048 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804206 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804360 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804504 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts\") pod \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\" (UID: \"89d7c14e-9ab5-45ec-8cae-6734847f9fa2\") " Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804587 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.804643 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.805354 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.805477 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.808932 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n" (OuterVolumeSpecName: "kube-api-access-f2s8n") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "kube-api-access-f2s8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.829421 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.840644 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts" (OuterVolumeSpecName: "scripts") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.843970 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.844541 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89d7c14e-9ab5-45ec-8cae-6734847f9fa2" (UID: "89d7c14e-9ab5-45ec-8cae-6734847f9fa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.907894 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.907939 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.907954 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2s8n\" (UniqueName: \"kubernetes.io/projected/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-kube-api-access-f2s8n\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.907967 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:28 crc kubenswrapper[4837]: I1001 10:06:28.907978 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89d7c14e-9ab5-45ec-8cae-6734847f9fa2-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.119351 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sn8dr" event={"ID":"67cf0e32-6d64-4bad-9b1b-3aee60fa713b","Type":"ContainerStarted","Data":"9aea42c3a90cd0f39d0567fee60b687d8e09390478b0256f71fca2d275aa9cff"} Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.119409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sn8dr" event={"ID":"67cf0e32-6d64-4bad-9b1b-3aee60fa713b","Type":"ContainerStarted","Data":"959a111dfd61991d6b911f8c2ea189761111ae2bc6bfddb8e1a0784a2492eefd"} Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.121998 4837 scope.go:117] "RemoveContainer" containerID="0bea40c3afc1c94782f1258dd226d84ae76dec151ed7e53f0c6730298a09d1ab" Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.122031 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2znrg" Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.158231 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-sn8dr" podStartSLOduration=2.158215975 podStartE2EDuration="2.158215975s" podCreationTimestamp="2025-10-01 10:06:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:06:29.142800543 +0000 UTC m=+10845.984408018" watchObservedRunningTime="2025-10-01 10:06:29.158215975 +0000 UTC m=+10845.999823420" Oct 01 10:06:29 crc kubenswrapper[4837]: I1001 10:06:29.833089 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d7c14e-9ab5-45ec-8cae-6734847f9fa2" path="/var/lib/kubelet/pods/89d7c14e-9ab5-45ec-8cae-6734847f9fa2/volumes" Oct 01 10:06:37 crc kubenswrapper[4837]: I1001 10:06:37.216578 4837 generic.go:334] "Generic (PLEG): container finished" podID="67cf0e32-6d64-4bad-9b1b-3aee60fa713b" containerID="9aea42c3a90cd0f39d0567fee60b687d8e09390478b0256f71fca2d275aa9cff" exitCode=0 Oct 01 10:06:37 crc kubenswrapper[4837]: I1001 10:06:37.216659 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sn8dr" event={"ID":"67cf0e32-6d64-4bad-9b1b-3aee60fa713b","Type":"ContainerDied","Data":"9aea42c3a90cd0f39d0567fee60b687d8e09390478b0256f71fca2d275aa9cff"} Oct 01 10:06:38 crc kubenswrapper[4837]: I1001 10:06:38.960824 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.007510 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-sn8dr"] Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.018105 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-sn8dr"] Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041243 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041309 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041418 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041467 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041588 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wtlv\" (UniqueName: \"kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041641 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.041684 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift\") pod \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\" (UID: \"67cf0e32-6d64-4bad-9b1b-3aee60fa713b\") " Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.042496 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.043026 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.062079 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv" (OuterVolumeSpecName: "kube-api-access-8wtlv") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "kube-api-access-8wtlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.081368 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts" (OuterVolumeSpecName: "scripts") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.085730 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.094631 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.097148 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "67cf0e32-6d64-4bad-9b1b-3aee60fa713b" (UID: "67cf0e32-6d64-4bad-9b1b-3aee60fa713b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.143966 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144004 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144017 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144027 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144041 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144055 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wtlv\" (UniqueName: \"kubernetes.io/projected/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-kube-api-access-8wtlv\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.144068 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67cf0e32-6d64-4bad-9b1b-3aee60fa713b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.240158 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="959a111dfd61991d6b911f8c2ea189761111ae2bc6bfddb8e1a0784a2492eefd" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.240193 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sn8dr" Oct 01 10:06:39 crc kubenswrapper[4837]: I1001 10:06:39.831568 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67cf0e32-6d64-4bad-9b1b-3aee60fa713b" path="/var/lib/kubelet/pods/67cf0e32-6d64-4bad-9b1b-3aee60fa713b/volumes" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.706005 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-sdkcr"] Oct 01 10:06:42 crc kubenswrapper[4837]: E1001 10:06:42.707009 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cf0e32-6d64-4bad-9b1b-3aee60fa713b" containerName="swift-ring-rebalance" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.707028 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cf0e32-6d64-4bad-9b1b-3aee60fa713b" containerName="swift-ring-rebalance" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.707351 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cf0e32-6d64-4bad-9b1b-3aee60fa713b" containerName="swift-ring-rebalance" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.708334 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.711106 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.711262 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.720926 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sdkcr"] Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.824634 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.827977 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.831143 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.831432 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.831554 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l972m\" (UniqueName: \"kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.831862 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.831987 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.832098 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.835508 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.852385 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937274 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937342 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937389 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937439 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937462 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l972m\" (UniqueName: \"kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937525 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937606 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937655 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937687 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.937736 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvprk\" (UniqueName: \"kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.938780 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.938960 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.939423 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.951389 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.951641 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.955756 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:42 crc kubenswrapper[4837]: I1001 10:06:42.957759 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l972m\" (UniqueName: \"kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m\") pod \"swift-ring-rebalance-debug-sdkcr\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.039803 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvprk\" (UniqueName: \"kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.039932 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.040012 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.040503 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.040523 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.047488 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.059670 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvprk\" (UniqueName: \"kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk\") pod \"community-operators-bnctn\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.153439 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.703444 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sdkcr"] Oct 01 10:06:43 crc kubenswrapper[4837]: I1001 10:06:43.843592 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:06:43 crc kubenswrapper[4837]: W1001 10:06:43.848968 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54afe01c_5290_45d6_8da3_81bfb1c20953.slice/crio-117962867ecb79454d99b627f806ea0085034398e6a4e6dfd0c64c32a19f45c4 WatchSource:0}: Error finding container 117962867ecb79454d99b627f806ea0085034398e6a4e6dfd0c64c32a19f45c4: Status 404 returned error can't find the container with id 117962867ecb79454d99b627f806ea0085034398e6a4e6dfd0c64c32a19f45c4 Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.351947 4837 generic.go:334] "Generic (PLEG): container finished" podID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerID="68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11" exitCode=0 Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.352021 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerDied","Data":"68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11"} Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.352294 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerStarted","Data":"117962867ecb79454d99b627f806ea0085034398e6a4e6dfd0c64c32a19f45c4"} Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.354291 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sdkcr" event={"ID":"a8c57092-0576-4e88-b0ef-371b78b906fc","Type":"ContainerStarted","Data":"51c5c631355bffda4e026d4521fd37cc4734b98057173875e57abca940d6724e"} Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.354317 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sdkcr" event={"ID":"a8c57092-0576-4e88-b0ef-371b78b906fc","Type":"ContainerStarted","Data":"da01d2f1faa6c0962352f41b81a9ecb65fcad61effc1cd292a6fe91c3e2c0e47"} Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.355778 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:06:44 crc kubenswrapper[4837]: I1001 10:06:44.396797 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-sdkcr" podStartSLOduration=2.396773712 podStartE2EDuration="2.396773712s" podCreationTimestamp="2025-10-01 10:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:06:44.395460439 +0000 UTC m=+10861.237067914" watchObservedRunningTime="2025-10-01 10:06:44.396773712 +0000 UTC m=+10861.238381197" Oct 01 10:06:45 crc kubenswrapper[4837]: I1001 10:06:45.366745 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerStarted","Data":"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270"} Oct 01 10:06:48 crc kubenswrapper[4837]: I1001 10:06:48.402486 4837 generic.go:334] "Generic (PLEG): container finished" podID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerID="9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270" exitCode=0 Oct 01 10:06:48 crc kubenswrapper[4837]: I1001 10:06:48.402559 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerDied","Data":"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270"} Oct 01 10:06:50 crc kubenswrapper[4837]: I1001 10:06:50.435045 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerStarted","Data":"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5"} Oct 01 10:06:50 crc kubenswrapper[4837]: I1001 10:06:50.461093 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bnctn" podStartSLOduration=3.73200237 podStartE2EDuration="8.461068718s" podCreationTimestamp="2025-10-01 10:06:42 +0000 UTC" firstStartedPulling="2025-10-01 10:06:44.35550941 +0000 UTC m=+10861.197116865" lastFinishedPulling="2025-10-01 10:06:49.084575758 +0000 UTC m=+10865.926183213" observedRunningTime="2025-10-01 10:06:50.456093855 +0000 UTC m=+10867.297701340" watchObservedRunningTime="2025-10-01 10:06:50.461068718 +0000 UTC m=+10867.302676193" Oct 01 10:06:53 crc kubenswrapper[4837]: I1001 10:06:53.153834 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:53 crc kubenswrapper[4837]: I1001 10:06:53.154262 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:06:54 crc kubenswrapper[4837]: I1001 10:06:54.208174 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bnctn" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="registry-server" probeResult="failure" output=< Oct 01 10:06:54 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:06:54 crc kubenswrapper[4837]: > Oct 01 10:06:54 crc kubenswrapper[4837]: I1001 10:06:54.495572 4837 generic.go:334] "Generic (PLEG): container finished" podID="a8c57092-0576-4e88-b0ef-371b78b906fc" containerID="51c5c631355bffda4e026d4521fd37cc4734b98057173875e57abca940d6724e" exitCode=0 Oct 01 10:06:54 crc kubenswrapper[4837]: I1001 10:06:54.495636 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sdkcr" event={"ID":"a8c57092-0576-4e88-b0ef-371b78b906fc","Type":"ContainerDied","Data":"51c5c631355bffda4e026d4521fd37cc4734b98057173875e57abca940d6724e"} Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.446302 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.504768 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-sdkcr"] Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.520435 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-sdkcr"] Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.530541 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da01d2f1faa6c0962352f41b81a9ecb65fcad61effc1cd292a6fe91c3e2c0e47" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.530638 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sdkcr" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.586536 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l972m\" (UniqueName: \"kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.586657 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.586816 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.586954 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.587000 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.587107 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.587153 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf\") pod \"a8c57092-0576-4e88-b0ef-371b78b906fc\" (UID: \"a8c57092-0576-4e88-b0ef-371b78b906fc\") " Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.624048 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m" (OuterVolumeSpecName: "kube-api-access-l972m") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "kube-api-access-l972m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.690755 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l972m\" (UniqueName: \"kubernetes.io/projected/a8c57092-0576-4e88-b0ef-371b78b906fc-kube-api-access-l972m\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.696229 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.733963 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts" (OuterVolumeSpecName: "scripts") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.757607 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.774792 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.774830 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.775441 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a8c57092-0576-4e88-b0ef-371b78b906fc" (UID: "a8c57092-0576-4e88-b0ef-371b78b906fc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793189 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793225 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793236 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a8c57092-0576-4e88-b0ef-371b78b906fc-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793245 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793255 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c57092-0576-4e88-b0ef-371b78b906fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:56 crc kubenswrapper[4837]: I1001 10:06:56.793266 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a8c57092-0576-4e88-b0ef-371b78b906fc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:57 crc kubenswrapper[4837]: I1001 10:06:57.835678 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c57092-0576-4e88-b0ef-371b78b906fc" path="/var/lib/kubelet/pods/a8c57092-0576-4e88-b0ef-371b78b906fc/volumes" Oct 01 10:07:03 crc kubenswrapper[4837]: I1001 10:07:03.206439 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:07:03 crc kubenswrapper[4837]: I1001 10:07:03.274149 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:07:03 crc kubenswrapper[4837]: I1001 10:07:03.454326 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:07:04 crc kubenswrapper[4837]: I1001 10:07:04.677973 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bnctn" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="registry-server" containerID="cri-o://407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5" gracePeriod=2 Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.347794 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.505609 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvprk\" (UniqueName: \"kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk\") pod \"54afe01c-5290-45d6-8da3-81bfb1c20953\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.505805 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities\") pod \"54afe01c-5290-45d6-8da3-81bfb1c20953\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.505904 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content\") pod \"54afe01c-5290-45d6-8da3-81bfb1c20953\" (UID: \"54afe01c-5290-45d6-8da3-81bfb1c20953\") " Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.508028 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities" (OuterVolumeSpecName: "utilities") pod "54afe01c-5290-45d6-8da3-81bfb1c20953" (UID: "54afe01c-5290-45d6-8da3-81bfb1c20953"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.511546 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk" (OuterVolumeSpecName: "kube-api-access-vvprk") pod "54afe01c-5290-45d6-8da3-81bfb1c20953" (UID: "54afe01c-5290-45d6-8da3-81bfb1c20953"). InnerVolumeSpecName "kube-api-access-vvprk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.556388 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54afe01c-5290-45d6-8da3-81bfb1c20953" (UID: "54afe01c-5290-45d6-8da3-81bfb1c20953"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.608830 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.609275 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54afe01c-5290-45d6-8da3-81bfb1c20953-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.609288 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvprk\" (UniqueName: \"kubernetes.io/projected/54afe01c-5290-45d6-8da3-81bfb1c20953-kube-api-access-vvprk\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.694854 4837 generic.go:334] "Generic (PLEG): container finished" podID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerID="407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5" exitCode=0 Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.694899 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerDied","Data":"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5"} Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.694927 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnctn" event={"ID":"54afe01c-5290-45d6-8da3-81bfb1c20953","Type":"ContainerDied","Data":"117962867ecb79454d99b627f806ea0085034398e6a4e6dfd0c64c32a19f45c4"} Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.694930 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnctn" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.694945 4837 scope.go:117] "RemoveContainer" containerID="407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.740371 4837 scope.go:117] "RemoveContainer" containerID="9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.741143 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.753666 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bnctn"] Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.768193 4837 scope.go:117] "RemoveContainer" containerID="68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.830412 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" path="/var/lib/kubelet/pods/54afe01c-5290-45d6-8da3-81bfb1c20953/volumes" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.839642 4837 scope.go:117] "RemoveContainer" containerID="407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5" Oct 01 10:07:05 crc kubenswrapper[4837]: E1001 10:07:05.840212 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5\": container with ID starting with 407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5 not found: ID does not exist" containerID="407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.840249 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5"} err="failed to get container status \"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5\": rpc error: code = NotFound desc = could not find container \"407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5\": container with ID starting with 407ea72fd281eebff8e3b741f744a437ccc55ae3e3d4c2597d93f353533477e5 not found: ID does not exist" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.840276 4837 scope.go:117] "RemoveContainer" containerID="9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270" Oct 01 10:07:05 crc kubenswrapper[4837]: E1001 10:07:05.840829 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270\": container with ID starting with 9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270 not found: ID does not exist" containerID="9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.840863 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270"} err="failed to get container status \"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270\": rpc error: code = NotFound desc = could not find container \"9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270\": container with ID starting with 9bbea438b98da8b913fdc2449d8734f243113f35301b13a220f5caa218c7f270 not found: ID does not exist" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.840878 4837 scope.go:117] "RemoveContainer" containerID="68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11" Oct 01 10:07:05 crc kubenswrapper[4837]: E1001 10:07:05.841200 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11\": container with ID starting with 68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11 not found: ID does not exist" containerID="68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11" Oct 01 10:07:05 crc kubenswrapper[4837]: I1001 10:07:05.841291 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11"} err="failed to get container status \"68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11\": rpc error: code = NotFound desc = could not find container \"68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11\": container with ID starting with 68169767a445632a26526cb87a133654c6f9fc84bc68c7dc80648c2636ebdb11 not found: ID does not exist" Oct 01 10:07:53 crc kubenswrapper[4837]: I1001 10:07:53.079089 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:07:53 crc kubenswrapper[4837]: I1001 10:07:53.079789 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.655216 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-zf64d"] Oct 01 10:07:56 crc kubenswrapper[4837]: E1001 10:07:56.656292 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="extract-utilities" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656307 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="extract-utilities" Oct 01 10:07:56 crc kubenswrapper[4837]: E1001 10:07:56.656330 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="extract-content" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656336 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="extract-content" Oct 01 10:07:56 crc kubenswrapper[4837]: E1001 10:07:56.656351 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c57092-0576-4e88-b0ef-371b78b906fc" containerName="swift-ring-rebalance" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656357 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c57092-0576-4e88-b0ef-371b78b906fc" containerName="swift-ring-rebalance" Oct 01 10:07:56 crc kubenswrapper[4837]: E1001 10:07:56.656382 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="registry-server" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656389 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="registry-server" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656668 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c57092-0576-4e88-b0ef-371b78b906fc" containerName="swift-ring-rebalance" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.656719 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="54afe01c-5290-45d6-8da3-81bfb1c20953" containerName="registry-server" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.657598 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.659717 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.659725 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.669478 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-zf64d"] Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.766515 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.766604 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.766930 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8bsc\" (UniqueName: \"kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.767278 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.767339 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.767455 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.767613 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869236 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869281 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869316 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869368 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869426 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869475 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869522 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8bsc\" (UniqueName: \"kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.869901 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.870398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.870465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.876118 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.876357 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.882522 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.887166 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8bsc\" (UniqueName: \"kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc\") pod \"swift-ring-rebalance-debug-zf64d\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:56 crc kubenswrapper[4837]: I1001 10:07:56.990967 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:07:57 crc kubenswrapper[4837]: I1001 10:07:57.502480 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-zf64d"] Oct 01 10:07:58 crc kubenswrapper[4837]: I1001 10:07:58.316753 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-zf64d" event={"ID":"13a57c29-95c7-4f15-9438-2f7297c69342","Type":"ContainerStarted","Data":"d5e780133c45c124f08390383be46362f9e411371e9f841a8f986d6cb9ee6b0d"} Oct 01 10:07:58 crc kubenswrapper[4837]: I1001 10:07:58.317481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-zf64d" event={"ID":"13a57c29-95c7-4f15-9438-2f7297c69342","Type":"ContainerStarted","Data":"d3ac9dda2e753dd78aa8c2547de9c186463ee1ebd38e5f6a52f074fa09e9c79a"} Oct 01 10:07:58 crc kubenswrapper[4837]: I1001 10:07:58.367137 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-zf64d" podStartSLOduration=2.367120209 podStartE2EDuration="2.367120209s" podCreationTimestamp="2025-10-01 10:07:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:07:58.351936795 +0000 UTC m=+10935.193544250" watchObservedRunningTime="2025-10-01 10:07:58.367120209 +0000 UTC m=+10935.208727664" Oct 01 10:08:06 crc kubenswrapper[4837]: I1001 10:08:06.415525 4837 generic.go:334] "Generic (PLEG): container finished" podID="13a57c29-95c7-4f15-9438-2f7297c69342" containerID="d5e780133c45c124f08390383be46362f9e411371e9f841a8f986d6cb9ee6b0d" exitCode=0 Oct 01 10:08:06 crc kubenswrapper[4837]: I1001 10:08:06.415585 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-zf64d" event={"ID":"13a57c29-95c7-4f15-9438-2f7297c69342","Type":"ContainerDied","Data":"d5e780133c45c124f08390383be46362f9e411371e9f841a8f986d6cb9ee6b0d"} Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.278754 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.344823 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-zf64d"] Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.360916 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-zf64d"] Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.421685 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.421811 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.421900 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.421943 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.422004 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8bsc\" (UniqueName: \"kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.422203 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.422897 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.423111 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.423194 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle\") pod \"13a57c29-95c7-4f15-9438-2f7297c69342\" (UID: \"13a57c29-95c7-4f15-9438-2f7297c69342\") " Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.424051 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/13a57c29-95c7-4f15-9438-2f7297c69342-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.424080 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.428143 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc" (OuterVolumeSpecName: "kube-api-access-g8bsc") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "kube-api-access-g8bsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.446863 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3ac9dda2e753dd78aa8c2547de9c186463ee1ebd38e5f6a52f074fa09e9c79a" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.446928 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-zf64d" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.468375 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.472297 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.492422 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts" (OuterVolumeSpecName: "scripts") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.499822 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "13a57c29-95c7-4f15-9438-2f7297c69342" (UID: "13a57c29-95c7-4f15-9438-2f7297c69342"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.526426 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.526660 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13a57c29-95c7-4f15-9438-2f7297c69342-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.526673 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.526705 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8bsc\" (UniqueName: \"kubernetes.io/projected/13a57c29-95c7-4f15-9438-2f7297c69342-kube-api-access-g8bsc\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:08 crc kubenswrapper[4837]: I1001 10:08:08.526720 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a57c29-95c7-4f15-9438-2f7297c69342-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:09 crc kubenswrapper[4837]: I1001 10:08:09.827058 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13a57c29-95c7-4f15-9438-2f7297c69342" path="/var/lib/kubelet/pods/13a57c29-95c7-4f15-9438-2f7297c69342/volumes" Oct 01 10:08:18 crc kubenswrapper[4837]: I1001 10:08:18.971286 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:18 crc kubenswrapper[4837]: E1001 10:08:18.972792 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a57c29-95c7-4f15-9438-2f7297c69342" containerName="swift-ring-rebalance" Oct 01 10:08:18 crc kubenswrapper[4837]: I1001 10:08:18.972810 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a57c29-95c7-4f15-9438-2f7297c69342" containerName="swift-ring-rebalance" Oct 01 10:08:18 crc kubenswrapper[4837]: I1001 10:08:18.973611 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="13a57c29-95c7-4f15-9438-2f7297c69342" containerName="swift-ring-rebalance" Oct 01 10:08:18 crc kubenswrapper[4837]: I1001 10:08:18.978502 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:18 crc kubenswrapper[4837]: I1001 10:08:18.991137 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.055530 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.055819 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw587\" (UniqueName: \"kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.055879 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.157615 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw587\" (UniqueName: \"kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.157679 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.157817 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.158397 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.158450 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.186723 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw587\" (UniqueName: \"kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587\") pod \"certified-operators-hd6zq\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:19 crc kubenswrapper[4837]: I1001 10:08:19.312086 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:20 crc kubenswrapper[4837]: I1001 10:08:20.190365 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:20 crc kubenswrapper[4837]: I1001 10:08:20.603177 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerID="6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963" exitCode=0 Oct 01 10:08:20 crc kubenswrapper[4837]: I1001 10:08:20.603227 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerDied","Data":"6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963"} Oct 01 10:08:20 crc kubenswrapper[4837]: I1001 10:08:20.603259 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerStarted","Data":"d1865f2c5754d0adc541563363f61c87bdedb00c0ce9a30fce9136aed9147c3d"} Oct 01 10:08:21 crc kubenswrapper[4837]: I1001 10:08:21.628663 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerStarted","Data":"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb"} Oct 01 10:08:23 crc kubenswrapper[4837]: I1001 10:08:23.079151 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:08:23 crc kubenswrapper[4837]: I1001 10:08:23.079568 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:08:23 crc kubenswrapper[4837]: I1001 10:08:23.663770 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerID="dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb" exitCode=0 Oct 01 10:08:23 crc kubenswrapper[4837]: I1001 10:08:23.663807 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerDied","Data":"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb"} Oct 01 10:08:24 crc kubenswrapper[4837]: I1001 10:08:24.678566 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerStarted","Data":"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0"} Oct 01 10:08:24 crc kubenswrapper[4837]: I1001 10:08:24.700454 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hd6zq" podStartSLOduration=3.211335472 podStartE2EDuration="6.700432737s" podCreationTimestamp="2025-10-01 10:08:18 +0000 UTC" firstStartedPulling="2025-10-01 10:08:20.60583398 +0000 UTC m=+10957.447441445" lastFinishedPulling="2025-10-01 10:08:24.094931235 +0000 UTC m=+10960.936538710" observedRunningTime="2025-10-01 10:08:24.699254438 +0000 UTC m=+10961.540861893" watchObservedRunningTime="2025-10-01 10:08:24.700432737 +0000 UTC m=+10961.542040192" Oct 01 10:08:29 crc kubenswrapper[4837]: I1001 10:08:29.312549 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:29 crc kubenswrapper[4837]: I1001 10:08:29.313178 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:29 crc kubenswrapper[4837]: I1001 10:08:29.364671 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:29 crc kubenswrapper[4837]: I1001 10:08:29.838800 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:29 crc kubenswrapper[4837]: I1001 10:08:29.907430 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:31 crc kubenswrapper[4837]: I1001 10:08:31.776254 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hd6zq" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="registry-server" containerID="cri-o://6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0" gracePeriod=2 Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.433635 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.604352 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities\") pod \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.604402 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content\") pod \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.604441 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw587\" (UniqueName: \"kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587\") pod \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\" (UID: \"c2e44525-5ba9-4ced-9a8c-3f2af3783472\") " Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.605195 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities" (OuterVolumeSpecName: "utilities") pod "c2e44525-5ba9-4ced-9a8c-3f2af3783472" (UID: "c2e44525-5ba9-4ced-9a8c-3f2af3783472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.615937 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587" (OuterVolumeSpecName: "kube-api-access-pw587") pod "c2e44525-5ba9-4ced-9a8c-3f2af3783472" (UID: "c2e44525-5ba9-4ced-9a8c-3f2af3783472"). InnerVolumeSpecName "kube-api-access-pw587". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.662592 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2e44525-5ba9-4ced-9a8c-3f2af3783472" (UID: "c2e44525-5ba9-4ced-9a8c-3f2af3783472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.707503 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.707801 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2e44525-5ba9-4ced-9a8c-3f2af3783472-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.707885 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw587\" (UniqueName: \"kubernetes.io/projected/c2e44525-5ba9-4ced-9a8c-3f2af3783472-kube-api-access-pw587\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.789772 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hd6zq" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.789683 4837 generic.go:334] "Generic (PLEG): container finished" podID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerID="6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0" exitCode=0 Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.789826 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerDied","Data":"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0"} Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.790131 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hd6zq" event={"ID":"c2e44525-5ba9-4ced-9a8c-3f2af3783472","Type":"ContainerDied","Data":"d1865f2c5754d0adc541563363f61c87bdedb00c0ce9a30fce9136aed9147c3d"} Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.790153 4837 scope.go:117] "RemoveContainer" containerID="6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.811820 4837 scope.go:117] "RemoveContainer" containerID="dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.838356 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.847833 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hd6zq"] Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.869702 4837 scope.go:117] "RemoveContainer" containerID="6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.899743 4837 scope.go:117] "RemoveContainer" containerID="6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0" Oct 01 10:08:32 crc kubenswrapper[4837]: E1001 10:08:32.900277 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0\": container with ID starting with 6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0 not found: ID does not exist" containerID="6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.900322 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0"} err="failed to get container status \"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0\": rpc error: code = NotFound desc = could not find container \"6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0\": container with ID starting with 6b7005511cbc0445cf8c0db30587f27824cebf88077749d395efe5b8115cc2e0 not found: ID does not exist" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.900349 4837 scope.go:117] "RemoveContainer" containerID="dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb" Oct 01 10:08:32 crc kubenswrapper[4837]: E1001 10:08:32.900890 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb\": container with ID starting with dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb not found: ID does not exist" containerID="dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.900943 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb"} err="failed to get container status \"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb\": rpc error: code = NotFound desc = could not find container \"dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb\": container with ID starting with dda0e5e28eadb64e6c3236c42d03cdba0d11a6b1fff930630c2830d4d7913fbb not found: ID does not exist" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.900976 4837 scope.go:117] "RemoveContainer" containerID="6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963" Oct 01 10:08:32 crc kubenswrapper[4837]: E1001 10:08:32.901295 4837 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963\": container with ID starting with 6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963 not found: ID does not exist" containerID="6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963" Oct 01 10:08:32 crc kubenswrapper[4837]: I1001 10:08:32.901322 4837 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963"} err="failed to get container status \"6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963\": rpc error: code = NotFound desc = could not find container \"6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963\": container with ID starting with 6a7cfad1060125acd3391926bd973159849b90f12f48912ef9539d49d8608963 not found: ID does not exist" Oct 01 10:08:33 crc kubenswrapper[4837]: I1001 10:08:33.836144 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" path="/var/lib/kubelet/pods/c2e44525-5ba9-4ced-9a8c-3f2af3783472/volumes" Oct 01 10:08:53 crc kubenswrapper[4837]: I1001 10:08:53.079794 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:08:53 crc kubenswrapper[4837]: I1001 10:08:53.080622 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:08:53 crc kubenswrapper[4837]: I1001 10:08:53.080680 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:08:53 crc kubenswrapper[4837]: I1001 10:08:53.082672 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:08:53 crc kubenswrapper[4837]: I1001 10:08:53.082761 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b" gracePeriod=600 Oct 01 10:08:54 crc kubenswrapper[4837]: I1001 10:08:54.083766 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b" exitCode=0 Oct 01 10:08:54 crc kubenswrapper[4837]: I1001 10:08:54.083828 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b"} Oct 01 10:08:54 crc kubenswrapper[4837]: I1001 10:08:54.084119 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a"} Oct 01 10:08:54 crc kubenswrapper[4837]: I1001 10:08:54.084147 4837 scope.go:117] "RemoveContainer" containerID="87c0cb63deeef1cdc983f76cae4fdfd940c284818d3e3737b3ea24cb3044c9ba" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.499897 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-rjdr8"] Oct 01 10:09:08 crc kubenswrapper[4837]: E1001 10:09:08.501832 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="extract-utilities" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.501924 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="extract-utilities" Oct 01 10:09:08 crc kubenswrapper[4837]: E1001 10:09:08.502048 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="registry-server" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.502118 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="registry-server" Oct 01 10:09:08 crc kubenswrapper[4837]: E1001 10:09:08.502208 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="extract-content" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.502272 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="extract-content" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.502560 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e44525-5ba9-4ced-9a8c-3f2af3783472" containerName="registry-server" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.503471 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.505896 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.506173 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.511851 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-rjdr8"] Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.604277 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.604364 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrgqj\" (UniqueName: \"kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.604435 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.604523 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.604635 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.606185 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.606398 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708501 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708587 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708618 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrgqj\" (UniqueName: \"kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708647 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708687 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708741 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.708779 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.709128 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.709670 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.710365 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.716391 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.716822 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.722620 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.729129 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrgqj\" (UniqueName: \"kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj\") pod \"swift-ring-rebalance-debug-rjdr8\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:08 crc kubenswrapper[4837]: I1001 10:09:08.863899 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:09 crc kubenswrapper[4837]: I1001 10:09:09.359151 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-rjdr8"] Oct 01 10:09:10 crc kubenswrapper[4837]: I1001 10:09:10.343345 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-rjdr8" event={"ID":"94128f59-ef7f-4cbf-8b73-7b642a2c300d","Type":"ContainerStarted","Data":"9423bfcadcba48b768e282dcbe1347a541f5870e591724c2eb37214c9c677123"} Oct 01 10:09:10 crc kubenswrapper[4837]: I1001 10:09:10.343880 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-rjdr8" event={"ID":"94128f59-ef7f-4cbf-8b73-7b642a2c300d","Type":"ContainerStarted","Data":"b726b3f3aa6a345b1dacba6974566297f0e3573db292e0d4b48cc3823b4fccba"} Oct 01 10:09:10 crc kubenswrapper[4837]: I1001 10:09:10.376862 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-rjdr8" podStartSLOduration=2.376834097 podStartE2EDuration="2.376834097s" podCreationTimestamp="2025-10-01 10:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:09:10.359081707 +0000 UTC m=+11007.200689182" watchObservedRunningTime="2025-10-01 10:09:10.376834097 +0000 UTC m=+11007.218441592" Oct 01 10:09:18 crc kubenswrapper[4837]: I1001 10:09:18.454532 4837 generic.go:334] "Generic (PLEG): container finished" podID="94128f59-ef7f-4cbf-8b73-7b642a2c300d" containerID="9423bfcadcba48b768e282dcbe1347a541f5870e591724c2eb37214c9c677123" exitCode=0 Oct 01 10:09:18 crc kubenswrapper[4837]: I1001 10:09:18.454625 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-rjdr8" event={"ID":"94128f59-ef7f-4cbf-8b73-7b642a2c300d","Type":"ContainerDied","Data":"9423bfcadcba48b768e282dcbe1347a541f5870e591724c2eb37214c9c677123"} Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.740657 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.790981 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-rjdr8"] Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.798606 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-rjdr8"] Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884031 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrgqj\" (UniqueName: \"kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884194 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884264 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884361 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884401 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884440 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.884481 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf\") pod \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\" (UID: \"94128f59-ef7f-4cbf-8b73-7b642a2c300d\") " Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.885283 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.885347 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.890637 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj" (OuterVolumeSpecName: "kube-api-access-wrgqj") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "kube-api-access-wrgqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.921747 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts" (OuterVolumeSpecName: "scripts") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.926959 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.929310 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.939547 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94128f59-ef7f-4cbf-8b73-7b642a2c300d" (UID: "94128f59-ef7f-4cbf-8b73-7b642a2c300d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.987892 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.987943 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94128f59-ef7f-4cbf-8b73-7b642a2c300d-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.987963 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.987980 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.988001 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/94128f59-ef7f-4cbf-8b73-7b642a2c300d-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.988023 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/94128f59-ef7f-4cbf-8b73-7b642a2c300d-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:20 crc kubenswrapper[4837]: I1001 10:09:20.988046 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrgqj\" (UniqueName: \"kubernetes.io/projected/94128f59-ef7f-4cbf-8b73-7b642a2c300d-kube-api-access-wrgqj\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:21 crc kubenswrapper[4837]: I1001 10:09:21.492984 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b726b3f3aa6a345b1dacba6974566297f0e3573db292e0d4b48cc3823b4fccba" Oct 01 10:09:21 crc kubenswrapper[4837]: I1001 10:09:21.493102 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rjdr8" Oct 01 10:09:21 crc kubenswrapper[4837]: I1001 10:09:21.830439 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94128f59-ef7f-4cbf-8b73-7b642a2c300d" path="/var/lib/kubelet/pods/94128f59-ef7f-4cbf-8b73-7b642a2c300d/volumes" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.973235 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-tjjn2"] Oct 01 10:10:20 crc kubenswrapper[4837]: E1001 10:10:20.974431 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94128f59-ef7f-4cbf-8b73-7b642a2c300d" containerName="swift-ring-rebalance" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.974450 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="94128f59-ef7f-4cbf-8b73-7b642a2c300d" containerName="swift-ring-rebalance" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.974785 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="94128f59-ef7f-4cbf-8b73-7b642a2c300d" containerName="swift-ring-rebalance" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.975741 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.979026 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:10:20 crc kubenswrapper[4837]: I1001 10:10:20.979507 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.005879 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-tjjn2"] Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064066 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smrd2\" (UniqueName: \"kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064176 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064284 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064343 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064389 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064463 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.064625 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.166621 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smrd2\" (UniqueName: \"kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.166669 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.166817 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.166859 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.166875 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.167039 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.167620 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.167751 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.167927 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.168329 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.175992 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.176000 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.187620 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.197595 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smrd2\" (UniqueName: \"kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2\") pod \"swift-ring-rebalance-debug-tjjn2\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.316937 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:21 crc kubenswrapper[4837]: I1001 10:10:21.995270 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-tjjn2"] Oct 01 10:10:22 crc kubenswrapper[4837]: I1001 10:10:22.172416 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-tjjn2" event={"ID":"cdf4b8d9-6805-455b-9e1b-93e37028cd44","Type":"ContainerStarted","Data":"2a13067711be326a039cdfa54b1dcdc85f337eab1509443bd54ee59576a3a634"} Oct 01 10:10:23 crc kubenswrapper[4837]: I1001 10:10:23.184629 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-tjjn2" event={"ID":"cdf4b8d9-6805-455b-9e1b-93e37028cd44","Type":"ContainerStarted","Data":"2ad01effb8b14a33d700434ac2b1590d1e786ff4f7d2410a8132949b690998f8"} Oct 01 10:10:23 crc kubenswrapper[4837]: I1001 10:10:23.224384 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-tjjn2" podStartSLOduration=3.224357303 podStartE2EDuration="3.224357303s" podCreationTimestamp="2025-10-01 10:10:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:10:23.20925354 +0000 UTC m=+11080.050861005" watchObservedRunningTime="2025-10-01 10:10:23.224357303 +0000 UTC m=+11080.065964768" Oct 01 10:10:33 crc kubenswrapper[4837]: I1001 10:10:33.308573 4837 generic.go:334] "Generic (PLEG): container finished" podID="cdf4b8d9-6805-455b-9e1b-93e37028cd44" containerID="2ad01effb8b14a33d700434ac2b1590d1e786ff4f7d2410a8132949b690998f8" exitCode=0 Oct 01 10:10:33 crc kubenswrapper[4837]: I1001 10:10:33.309363 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-tjjn2" event={"ID":"cdf4b8d9-6805-455b-9e1b-93e37028cd44","Type":"ContainerDied","Data":"2ad01effb8b14a33d700434ac2b1590d1e786ff4f7d2410a8132949b690998f8"} Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.651012 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.724071 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-tjjn2"] Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.733758 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-tjjn2"] Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739424 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smrd2\" (UniqueName: \"kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739484 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739552 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739688 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739758 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739838 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.739881 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift\") pod \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\" (UID: \"cdf4b8d9-6805-455b-9e1b-93e37028cd44\") " Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.741007 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.742817 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.748633 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2" (OuterVolumeSpecName: "kube-api-access-smrd2") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "kube-api-access-smrd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.775266 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.777018 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.782781 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts" (OuterVolumeSpecName: "scripts") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.787863 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cdf4b8d9-6805-455b-9e1b-93e37028cd44" (UID: "cdf4b8d9-6805-455b-9e1b-93e37028cd44"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.828879 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf4b8d9-6805-455b-9e1b-93e37028cd44" path="/var/lib/kubelet/pods/cdf4b8d9-6805-455b-9e1b-93e37028cd44/volumes" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842673 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cdf4b8d9-6805-455b-9e1b-93e37028cd44-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842710 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smrd2\" (UniqueName: \"kubernetes.io/projected/cdf4b8d9-6805-455b-9e1b-93e37028cd44-kube-api-access-smrd2\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842719 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842729 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842737 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842746 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf4b8d9-6805-455b-9e1b-93e37028cd44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:35 crc kubenswrapper[4837]: I1001 10:10:35.842754 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cdf4b8d9-6805-455b-9e1b-93e37028cd44-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:36 crc kubenswrapper[4837]: I1001 10:10:36.351584 4837 scope.go:117] "RemoveContainer" containerID="2ad01effb8b14a33d700434ac2b1590d1e786ff4f7d2410a8132949b690998f8" Oct 01 10:10:36 crc kubenswrapper[4837]: I1001 10:10:36.351677 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-tjjn2" Oct 01 10:10:53 crc kubenswrapper[4837]: I1001 10:10:53.079080 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:10:53 crc kubenswrapper[4837]: I1001 10:10:53.079592 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.619680 4837 scope.go:117] "RemoveContainer" containerID="c8cf2bf651fae65a77b6c46a6bd79ec83f94f398ebdf6c5c5ef707903621ce36" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.972037 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:17 crc kubenswrapper[4837]: E1001 10:11:17.972664 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf4b8d9-6805-455b-9e1b-93e37028cd44" containerName="swift-ring-rebalance" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.972684 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf4b8d9-6805-455b-9e1b-93e37028cd44" containerName="swift-ring-rebalance" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.973072 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf4b8d9-6805-455b-9e1b-93e37028cd44" containerName="swift-ring-rebalance" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.975117 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:17 crc kubenswrapper[4837]: I1001 10:11:17.990461 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.133784 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.134159 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt5bq\" (UniqueName: \"kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.134291 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.236615 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.236747 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt5bq\" (UniqueName: \"kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.236775 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.237341 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.237538 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.273447 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt5bq\" (UniqueName: \"kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq\") pod \"redhat-marketplace-26z55\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:18 crc kubenswrapper[4837]: I1001 10:11:18.298605 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:19 crc kubenswrapper[4837]: I1001 10:11:19.075419 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:19 crc kubenswrapper[4837]: I1001 10:11:19.950503 4837 generic.go:334] "Generic (PLEG): container finished" podID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerID="6c703d6a2bed3914ded52cce65bdec85d6ac326484bac08781fbc046f063d801" exitCode=0 Oct 01 10:11:19 crc kubenswrapper[4837]: I1001 10:11:19.950578 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerDied","Data":"6c703d6a2bed3914ded52cce65bdec85d6ac326484bac08781fbc046f063d801"} Oct 01 10:11:19 crc kubenswrapper[4837]: I1001 10:11:19.951703 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerStarted","Data":"8eb8bffb3f44290bbad5d35f09f821499712086020356a2b716404fd91944e88"} Oct 01 10:11:20 crc kubenswrapper[4837]: I1001 10:11:20.967939 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerStarted","Data":"a6449a00b0d32456a2b3ce274dd8c8f25752e56f2818fa1cde582a1f0ad45c05"} Oct 01 10:11:21 crc kubenswrapper[4837]: I1001 10:11:21.997788 4837 generic.go:334] "Generic (PLEG): container finished" podID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerID="a6449a00b0d32456a2b3ce274dd8c8f25752e56f2818fa1cde582a1f0ad45c05" exitCode=0 Oct 01 10:11:21 crc kubenswrapper[4837]: I1001 10:11:21.997913 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerDied","Data":"a6449a00b0d32456a2b3ce274dd8c8f25752e56f2818fa1cde582a1f0ad45c05"} Oct 01 10:11:23 crc kubenswrapper[4837]: I1001 10:11:23.013651 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerStarted","Data":"c0b25d858e7ce8ec7a2746dfd80c30033d382c514bed7ae1b72ce0c4be262e0a"} Oct 01 10:11:23 crc kubenswrapper[4837]: I1001 10:11:23.038631 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-26z55" podStartSLOduration=3.36082376 podStartE2EDuration="6.03860737s" podCreationTimestamp="2025-10-01 10:11:17 +0000 UTC" firstStartedPulling="2025-10-01 10:11:19.953844229 +0000 UTC m=+11136.795451684" lastFinishedPulling="2025-10-01 10:11:22.631627799 +0000 UTC m=+11139.473235294" observedRunningTime="2025-10-01 10:11:23.033234047 +0000 UTC m=+11139.874841502" watchObservedRunningTime="2025-10-01 10:11:23.03860737 +0000 UTC m=+11139.880214845" Oct 01 10:11:23 crc kubenswrapper[4837]: I1001 10:11:23.079558 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:11:23 crc kubenswrapper[4837]: I1001 10:11:23.079610 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:11:28 crc kubenswrapper[4837]: I1001 10:11:28.299655 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:28 crc kubenswrapper[4837]: I1001 10:11:28.300204 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:28 crc kubenswrapper[4837]: I1001 10:11:28.380376 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:29 crc kubenswrapper[4837]: I1001 10:11:29.165546 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:29 crc kubenswrapper[4837]: I1001 10:11:29.229618 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:31 crc kubenswrapper[4837]: I1001 10:11:31.110075 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-26z55" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="registry-server" containerID="cri-o://c0b25d858e7ce8ec7a2746dfd80c30033d382c514bed7ae1b72ce0c4be262e0a" gracePeriod=2 Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.128302 4837 generic.go:334] "Generic (PLEG): container finished" podID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerID="c0b25d858e7ce8ec7a2746dfd80c30033d382c514bed7ae1b72ce0c4be262e0a" exitCode=0 Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.128348 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerDied","Data":"c0b25d858e7ce8ec7a2746dfd80c30033d382c514bed7ae1b72ce0c4be262e0a"} Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.227452 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.345749 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content\") pod \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.345893 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities\") pod \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.345943 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt5bq\" (UniqueName: \"kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq\") pod \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\" (UID: \"5d24010c-3c3e-4a35-8599-6ac98c5afe64\") " Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.346667 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities" (OuterVolumeSpecName: "utilities") pod "5d24010c-3c3e-4a35-8599-6ac98c5afe64" (UID: "5d24010c-3c3e-4a35-8599-6ac98c5afe64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.352996 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq" (OuterVolumeSpecName: "kube-api-access-qt5bq") pod "5d24010c-3c3e-4a35-8599-6ac98c5afe64" (UID: "5d24010c-3c3e-4a35-8599-6ac98c5afe64"). InnerVolumeSpecName "kube-api-access-qt5bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.362165 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d24010c-3c3e-4a35-8599-6ac98c5afe64" (UID: "5d24010c-3c3e-4a35-8599-6ac98c5afe64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.448483 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.448518 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d24010c-3c3e-4a35-8599-6ac98c5afe64-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:32 crc kubenswrapper[4837]: I1001 10:11:32.448530 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt5bq\" (UniqueName: \"kubernetes.io/projected/5d24010c-3c3e-4a35-8599-6ac98c5afe64-kube-api-access-qt5bq\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.149400 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26z55" event={"ID":"5d24010c-3c3e-4a35-8599-6ac98c5afe64","Type":"ContainerDied","Data":"8eb8bffb3f44290bbad5d35f09f821499712086020356a2b716404fd91944e88"} Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.149461 4837 scope.go:117] "RemoveContainer" containerID="c0b25d858e7ce8ec7a2746dfd80c30033d382c514bed7ae1b72ce0c4be262e0a" Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.149502 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26z55" Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.205661 4837 scope.go:117] "RemoveContainer" containerID="a6449a00b0d32456a2b3ce274dd8c8f25752e56f2818fa1cde582a1f0ad45c05" Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.215575 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.231032 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-26z55"] Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.245388 4837 scope.go:117] "RemoveContainer" containerID="6c703d6a2bed3914ded52cce65bdec85d6ac326484bac08781fbc046f063d801" Oct 01 10:11:33 crc kubenswrapper[4837]: I1001 10:11:33.833310 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" path="/var/lib/kubelet/pods/5d24010c-3c3e-4a35-8599-6ac98c5afe64/volumes" Oct 01 10:11:35 crc kubenswrapper[4837]: I1001 10:11:35.999330 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-9n7fg"] Oct 01 10:11:36 crc kubenswrapper[4837]: E1001 10:11:36.000515 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="extract-content" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.000540 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="extract-content" Oct 01 10:11:36 crc kubenswrapper[4837]: E1001 10:11:36.000575 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="extract-utilities" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.000589 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="extract-utilities" Oct 01 10:11:36 crc kubenswrapper[4837]: E1001 10:11:36.000633 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="registry-server" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.000646 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="registry-server" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.001046 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d24010c-3c3e-4a35-8599-6ac98c5afe64" containerName="registry-server" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.002316 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.005602 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.005706 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.014895 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-9n7fg"] Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028013 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028350 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggb5r\" (UniqueName: \"kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028493 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028790 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028859 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.028954 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.130532 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggb5r\" (UniqueName: \"kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.130729 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.130779 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.130916 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.130957 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.131021 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.131124 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.131870 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.132101 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.132420 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.136860 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.137108 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.137934 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.152443 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggb5r\" (UniqueName: \"kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r\") pod \"swift-ring-rebalance-debug-9n7fg\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:36 crc kubenswrapper[4837]: I1001 10:11:36.344448 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:37 crc kubenswrapper[4837]: I1001 10:11:37.062075 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-9n7fg"] Oct 01 10:11:37 crc kubenswrapper[4837]: W1001 10:11:37.062440 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e16cdb0_3dec_454a_9abe_c59c666c6a0f.slice/crio-fd7c5c4881755f6715a2038e9f05286f9f1ae764846c5f439f4d7d86cc31bfb2 WatchSource:0}: Error finding container fd7c5c4881755f6715a2038e9f05286f9f1ae764846c5f439f4d7d86cc31bfb2: Status 404 returned error can't find the container with id fd7c5c4881755f6715a2038e9f05286f9f1ae764846c5f439f4d7d86cc31bfb2 Oct 01 10:11:37 crc kubenswrapper[4837]: I1001 10:11:37.200034 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9n7fg" event={"ID":"6e16cdb0-3dec-454a-9abe-c59c666c6a0f","Type":"ContainerStarted","Data":"fd7c5c4881755f6715a2038e9f05286f9f1ae764846c5f439f4d7d86cc31bfb2"} Oct 01 10:11:38 crc kubenswrapper[4837]: I1001 10:11:38.214054 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9n7fg" event={"ID":"6e16cdb0-3dec-454a-9abe-c59c666c6a0f","Type":"ContainerStarted","Data":"637b41e73303dc1610abaed84f5f841bf46ca6467531cd6b27b741e7e7b6efe5"} Oct 01 10:11:38 crc kubenswrapper[4837]: I1001 10:11:38.240917 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-9n7fg" podStartSLOduration=3.240891759 podStartE2EDuration="3.240891759s" podCreationTimestamp="2025-10-01 10:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:11:38.228656016 +0000 UTC m=+11155.070263491" watchObservedRunningTime="2025-10-01 10:11:38.240891759 +0000 UTC m=+11155.082499214" Oct 01 10:11:46 crc kubenswrapper[4837]: I1001 10:11:46.324184 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e16cdb0-3dec-454a-9abe-c59c666c6a0f" containerID="637b41e73303dc1610abaed84f5f841bf46ca6467531cd6b27b741e7e7b6efe5" exitCode=0 Oct 01 10:11:46 crc kubenswrapper[4837]: I1001 10:11:46.324253 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-9n7fg" event={"ID":"6e16cdb0-3dec-454a-9abe-c59c666c6a0f","Type":"ContainerDied","Data":"637b41e73303dc1610abaed84f5f841bf46ca6467531cd6b27b741e7e7b6efe5"} Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.016756 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.083573 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-9n7fg"] Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.094092 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-9n7fg"] Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.173888 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.173999 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggb5r\" (UniqueName: \"kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.174122 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.174248 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.174320 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.175304 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.175386 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices\") pod \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\" (UID: \"6e16cdb0-3dec-454a-9abe-c59c666c6a0f\") " Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.174832 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.176607 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.183444 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r" (OuterVolumeSpecName: "kube-api-access-ggb5r") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "kube-api-access-ggb5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.220076 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.223271 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.225337 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts" (OuterVolumeSpecName: "scripts") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.237538 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e16cdb0-3dec-454a-9abe-c59c666c6a0f" (UID: "6e16cdb0-3dec-454a-9abe-c59c666c6a0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278554 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278598 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278609 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278620 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278629 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggb5r\" (UniqueName: \"kubernetes.io/projected/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-kube-api-access-ggb5r\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278638 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.278646 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e16cdb0-3dec-454a-9abe-c59c666c6a0f-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.361766 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd7c5c4881755f6715a2038e9f05286f9f1ae764846c5f439f4d7d86cc31bfb2" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.361809 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-9n7fg" Oct 01 10:11:49 crc kubenswrapper[4837]: I1001 10:11:49.835663 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e16cdb0-3dec-454a-9abe-c59c666c6a0f" path="/var/lib/kubelet/pods/6e16cdb0-3dec-454a-9abe-c59c666c6a0f/volumes" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.079098 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.079731 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.079786 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.080619 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.080709 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" gracePeriod=600 Oct 01 10:11:53 crc kubenswrapper[4837]: E1001 10:11:53.206475 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.428979 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" exitCode=0 Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.429010 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a"} Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.429060 4837 scope.go:117] "RemoveContainer" containerID="0b23b1221cce08460cb877edf5c9093f7699a2d459f3a091c0106824d89d7e4b" Oct 01 10:11:53 crc kubenswrapper[4837]: I1001 10:11:53.429870 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:11:53 crc kubenswrapper[4837]: E1001 10:11:53.430339 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:12:08 crc kubenswrapper[4837]: I1001 10:12:08.818084 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:12:08 crc kubenswrapper[4837]: E1001 10:12:08.819120 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.286874 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:17 crc kubenswrapper[4837]: E1001 10:12:17.287749 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e16cdb0-3dec-454a-9abe-c59c666c6a0f" containerName="swift-ring-rebalance" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.287762 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e16cdb0-3dec-454a-9abe-c59c666c6a0f" containerName="swift-ring-rebalance" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.288122 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e16cdb0-3dec-454a-9abe-c59c666c6a0f" containerName="swift-ring-rebalance" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.290901 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.304058 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.358558 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.358615 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.358856 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ngnp\" (UniqueName: \"kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.460197 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.460255 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.460398 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ngnp\" (UniqueName: \"kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.461290 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.461554 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.479853 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ngnp\" (UniqueName: \"kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp\") pod \"redhat-operators-7cj88\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:17 crc kubenswrapper[4837]: I1001 10:12:17.627289 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:18 crc kubenswrapper[4837]: I1001 10:12:18.414289 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:18 crc kubenswrapper[4837]: I1001 10:12:18.762964 4837 generic.go:334] "Generic (PLEG): container finished" podID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerID="e093d9e0095a9c79e82b782f47add765671a4ccc4875a58f62c2b70e76e6eecb" exitCode=0 Oct 01 10:12:18 crc kubenswrapper[4837]: I1001 10:12:18.763020 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerDied","Data":"e093d9e0095a9c79e82b782f47add765671a4ccc4875a58f62c2b70e76e6eecb"} Oct 01 10:12:18 crc kubenswrapper[4837]: I1001 10:12:18.763046 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerStarted","Data":"b497b296de3fa0d3b1d223afc160332241aafae51d617472e2a504fd0077f2c7"} Oct 01 10:12:18 crc kubenswrapper[4837]: I1001 10:12:18.765307 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:12:19 crc kubenswrapper[4837]: I1001 10:12:19.779001 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerStarted","Data":"ae1efbd7f2bf4355976c6b5db4ed2d891a8571256ded569a36ae15fea0ac7c2b"} Oct 01 10:12:21 crc kubenswrapper[4837]: I1001 10:12:21.816388 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:12:21 crc kubenswrapper[4837]: E1001 10:12:21.818349 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:12:24 crc kubenswrapper[4837]: I1001 10:12:24.844847 4837 generic.go:334] "Generic (PLEG): container finished" podID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerID="ae1efbd7f2bf4355976c6b5db4ed2d891a8571256ded569a36ae15fea0ac7c2b" exitCode=0 Oct 01 10:12:24 crc kubenswrapper[4837]: I1001 10:12:24.844937 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerDied","Data":"ae1efbd7f2bf4355976c6b5db4ed2d891a8571256ded569a36ae15fea0ac7c2b"} Oct 01 10:12:25 crc kubenswrapper[4837]: I1001 10:12:25.859057 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerStarted","Data":"ce19fd4dd64bff9c543ad1ff36704c9dd80a3af4b5364e5b1ea70a120f18c043"} Oct 01 10:12:25 crc kubenswrapper[4837]: I1001 10:12:25.881784 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7cj88" podStartSLOduration=2.2360794840000002 podStartE2EDuration="8.881762787s" podCreationTimestamp="2025-10-01 10:12:17 +0000 UTC" firstStartedPulling="2025-10-01 10:12:18.765019968 +0000 UTC m=+11195.606627423" lastFinishedPulling="2025-10-01 10:12:25.410703271 +0000 UTC m=+11202.252310726" observedRunningTime="2025-10-01 10:12:25.878130846 +0000 UTC m=+11202.719738311" watchObservedRunningTime="2025-10-01 10:12:25.881762787 +0000 UTC m=+11202.723370252" Oct 01 10:12:27 crc kubenswrapper[4837]: I1001 10:12:27.628257 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:27 crc kubenswrapper[4837]: I1001 10:12:27.628310 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:28 crc kubenswrapper[4837]: I1001 10:12:28.687306 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7cj88" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="registry-server" probeResult="failure" output=< Oct 01 10:12:28 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:12:28 crc kubenswrapper[4837]: > Oct 01 10:12:34 crc kubenswrapper[4837]: I1001 10:12:34.816877 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:12:34 crc kubenswrapper[4837]: E1001 10:12:34.817654 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:12:37 crc kubenswrapper[4837]: I1001 10:12:37.690974 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:37 crc kubenswrapper[4837]: I1001 10:12:37.768662 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:37 crc kubenswrapper[4837]: I1001 10:12:37.932104 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:39 crc kubenswrapper[4837]: I1001 10:12:39.037706 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7cj88" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="registry-server" containerID="cri-o://ce19fd4dd64bff9c543ad1ff36704c9dd80a3af4b5364e5b1ea70a120f18c043" gracePeriod=2 Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.049345 4837 generic.go:334] "Generic (PLEG): container finished" podID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerID="ce19fd4dd64bff9c543ad1ff36704c9dd80a3af4b5364e5b1ea70a120f18c043" exitCode=0 Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.049542 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerDied","Data":"ce19fd4dd64bff9c543ad1ff36704c9dd80a3af4b5364e5b1ea70a120f18c043"} Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.505964 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.563190 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities\") pod \"eee697c4-ee04-4692-921b-e0dd98b114ae\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.563419 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ngnp\" (UniqueName: \"kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp\") pod \"eee697c4-ee04-4692-921b-e0dd98b114ae\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.563497 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content\") pod \"eee697c4-ee04-4692-921b-e0dd98b114ae\" (UID: \"eee697c4-ee04-4692-921b-e0dd98b114ae\") " Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.564453 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities" (OuterVolumeSpecName: "utilities") pod "eee697c4-ee04-4692-921b-e0dd98b114ae" (UID: "eee697c4-ee04-4692-921b-e0dd98b114ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.569807 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp" (OuterVolumeSpecName: "kube-api-access-8ngnp") pod "eee697c4-ee04-4692-921b-e0dd98b114ae" (UID: "eee697c4-ee04-4692-921b-e0dd98b114ae"). InnerVolumeSpecName "kube-api-access-8ngnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.666336 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.666380 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ngnp\" (UniqueName: \"kubernetes.io/projected/eee697c4-ee04-4692-921b-e0dd98b114ae-kube-api-access-8ngnp\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.674359 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eee697c4-ee04-4692-921b-e0dd98b114ae" (UID: "eee697c4-ee04-4692-921b-e0dd98b114ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:12:40 crc kubenswrapper[4837]: I1001 10:12:40.768938 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee697c4-ee04-4692-921b-e0dd98b114ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.065158 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cj88" event={"ID":"eee697c4-ee04-4692-921b-e0dd98b114ae","Type":"ContainerDied","Data":"b497b296de3fa0d3b1d223afc160332241aafae51d617472e2a504fd0077f2c7"} Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.065218 4837 scope.go:117] "RemoveContainer" containerID="ce19fd4dd64bff9c543ad1ff36704c9dd80a3af4b5364e5b1ea70a120f18c043" Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.065258 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cj88" Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.111968 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.121586 4837 scope.go:117] "RemoveContainer" containerID="ae1efbd7f2bf4355976c6b5db4ed2d891a8571256ded569a36ae15fea0ac7c2b" Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.134516 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7cj88"] Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.160492 4837 scope.go:117] "RemoveContainer" containerID="e093d9e0095a9c79e82b782f47add765671a4ccc4875a58f62c2b70e76e6eecb" Oct 01 10:12:41 crc kubenswrapper[4837]: I1001 10:12:41.830362 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" path="/var/lib/kubelet/pods/eee697c4-ee04-4692-921b-e0dd98b114ae/volumes" Oct 01 10:12:46 crc kubenswrapper[4837]: I1001 10:12:46.816469 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:12:46 crc kubenswrapper[4837]: E1001 10:12:46.817559 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.263673 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-xrqx7"] Oct 01 10:12:49 crc kubenswrapper[4837]: E1001 10:12:49.264867 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="registry-server" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.264894 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="registry-server" Oct 01 10:12:49 crc kubenswrapper[4837]: E1001 10:12:49.264961 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="extract-utilities" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.264974 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="extract-utilities" Oct 01 10:12:49 crc kubenswrapper[4837]: E1001 10:12:49.265005 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="extract-content" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.265020 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="extract-content" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.265456 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee697c4-ee04-4692-921b-e0dd98b114ae" containerName="registry-server" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.266818 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.269681 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.274104 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.275887 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-xrqx7"] Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350403 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350478 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350511 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tttk\" (UniqueName: \"kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350571 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350661 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.350833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.351108 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453429 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tttk\" (UniqueName: \"kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453534 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453555 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453585 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453640 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453724 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.453776 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.454129 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.454484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.455327 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.459380 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.459412 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.459731 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.472116 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tttk\" (UniqueName: \"kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk\") pod \"swift-ring-rebalance-debug-xrqx7\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:49 crc kubenswrapper[4837]: I1001 10:12:49.607154 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:12:50 crc kubenswrapper[4837]: I1001 10:12:50.335652 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-xrqx7"] Oct 01 10:12:51 crc kubenswrapper[4837]: I1001 10:12:51.189875 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xrqx7" event={"ID":"249a8d02-e5d6-4353-bf1d-541058a5da55","Type":"ContainerStarted","Data":"482d0a8fe52679d2c89dbdb4cbef1b60bfee2b4b19d8276bebb72ceb7be21e08"} Oct 01 10:12:51 crc kubenswrapper[4837]: I1001 10:12:51.190666 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xrqx7" event={"ID":"249a8d02-e5d6-4353-bf1d-541058a5da55","Type":"ContainerStarted","Data":"93707c1fa9b5872b55c9baaccb1702cf6983de55f289bf2907c7023c1436ade4"} Oct 01 10:12:51 crc kubenswrapper[4837]: I1001 10:12:51.223479 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-xrqx7" podStartSLOduration=2.223452337 podStartE2EDuration="2.223452337s" podCreationTimestamp="2025-10-01 10:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:12:51.208769734 +0000 UTC m=+11228.050377229" watchObservedRunningTime="2025-10-01 10:12:51.223452337 +0000 UTC m=+11228.065059822" Oct 01 10:12:57 crc kubenswrapper[4837]: I1001 10:12:57.817049 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:12:57 crc kubenswrapper[4837]: E1001 10:12:57.817927 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:13:00 crc kubenswrapper[4837]: I1001 10:13:00.294970 4837 generic.go:334] "Generic (PLEG): container finished" podID="249a8d02-e5d6-4353-bf1d-541058a5da55" containerID="482d0a8fe52679d2c89dbdb4cbef1b60bfee2b4b19d8276bebb72ceb7be21e08" exitCode=0 Oct 01 10:13:00 crc kubenswrapper[4837]: I1001 10:13:00.295095 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xrqx7" event={"ID":"249a8d02-e5d6-4353-bf1d-541058a5da55","Type":"ContainerDied","Data":"482d0a8fe52679d2c89dbdb4cbef1b60bfee2b4b19d8276bebb72ceb7be21e08"} Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.931372 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.964873 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.964928 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.964991 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.965074 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tttk\" (UniqueName: \"kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.965144 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.966409 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.966500 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices\") pod \"249a8d02-e5d6-4353-bf1d-541058a5da55\" (UID: \"249a8d02-e5d6-4353-bf1d-541058a5da55\") " Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.967286 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.967385 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.972577 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk" (OuterVolumeSpecName: "kube-api-access-5tttk") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "kube-api-access-5tttk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:13:02 crc kubenswrapper[4837]: I1001 10:13:02.994353 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-xrqx7"] Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.011609 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.020358 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.021738 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-xrqx7"] Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.033274 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts" (OuterVolumeSpecName: "scripts") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.053011 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "249a8d02-e5d6-4353-bf1d-541058a5da55" (UID: "249a8d02-e5d6-4353-bf1d-541058a5da55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069222 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069256 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/249a8d02-e5d6-4353-bf1d-541058a5da55-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069266 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069275 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069284 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tttk\" (UniqueName: \"kubernetes.io/projected/249a8d02-e5d6-4353-bf1d-541058a5da55-kube-api-access-5tttk\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069295 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/249a8d02-e5d6-4353-bf1d-541058a5da55-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.069303 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/249a8d02-e5d6-4353-bf1d-541058a5da55-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.346718 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93707c1fa9b5872b55c9baaccb1702cf6983de55f289bf2907c7023c1436ade4" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.346854 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xrqx7" Oct 01 10:13:03 crc kubenswrapper[4837]: I1001 10:13:03.827968 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="249a8d02-e5d6-4353-bf1d-541058a5da55" path="/var/lib/kubelet/pods/249a8d02-e5d6-4353-bf1d-541058a5da55/volumes" Oct 01 10:13:11 crc kubenswrapper[4837]: I1001 10:13:11.816608 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:13:11 crc kubenswrapper[4837]: E1001 10:13:11.817570 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:13:17 crc kubenswrapper[4837]: I1001 10:13:17.805767 4837 scope.go:117] "RemoveContainer" containerID="9aea42c3a90cd0f39d0567fee60b687d8e09390478b0256f71fca2d275aa9cff" Oct 01 10:13:17 crc kubenswrapper[4837]: I1001 10:13:17.882283 4837 scope.go:117] "RemoveContainer" containerID="51c5c631355bffda4e026d4521fd37cc4734b98057173875e57abca940d6724e" Oct 01 10:13:22 crc kubenswrapper[4837]: I1001 10:13:22.816068 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:13:22 crc kubenswrapper[4837]: E1001 10:13:22.816912 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:13:34 crc kubenswrapper[4837]: I1001 10:13:34.816158 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:13:34 crc kubenswrapper[4837]: E1001 10:13:34.816865 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:13:46 crc kubenswrapper[4837]: I1001 10:13:46.816597 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:13:46 crc kubenswrapper[4837]: E1001 10:13:46.817771 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:13:59 crc kubenswrapper[4837]: I1001 10:13:59.816998 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:13:59 crc kubenswrapper[4837]: E1001 10:13:59.818116 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.180840 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-sk5cn"] Oct 01 10:14:03 crc kubenswrapper[4837]: E1001 10:14:03.182195 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="249a8d02-e5d6-4353-bf1d-541058a5da55" containerName="swift-ring-rebalance" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.182228 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="249a8d02-e5d6-4353-bf1d-541058a5da55" containerName="swift-ring-rebalance" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.182587 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="249a8d02-e5d6-4353-bf1d-541058a5da55" containerName="swift-ring-rebalance" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.184221 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.186507 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.194389 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sk5cn"] Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.201539 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.338852 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.339924 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.340029 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.340139 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.340193 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.340244 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq2ql\" (UniqueName: \"kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.340291 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.442943 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443130 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443219 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443270 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443310 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443353 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq2ql\" (UniqueName: \"kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443397 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.443579 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.444351 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.444673 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.452336 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.452559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.452657 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.462316 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq2ql\" (UniqueName: \"kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql\") pod \"swift-ring-rebalance-debug-sk5cn\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:03 crc kubenswrapper[4837]: I1001 10:14:03.514619 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:04 crc kubenswrapper[4837]: I1001 10:14:04.495098 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-sk5cn"] Oct 01 10:14:04 crc kubenswrapper[4837]: W1001 10:14:04.507684 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3bc9902_a999_4ebd_a6d9_23892a188dc5.slice/crio-215c99815631f5dcb13b0cd7e1eea379d093ed6f870bd4d4318ab794e6c58859 WatchSource:0}: Error finding container 215c99815631f5dcb13b0cd7e1eea379d093ed6f870bd4d4318ab794e6c58859: Status 404 returned error can't find the container with id 215c99815631f5dcb13b0cd7e1eea379d093ed6f870bd4d4318ab794e6c58859 Oct 01 10:14:05 crc kubenswrapper[4837]: I1001 10:14:05.139015 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sk5cn" event={"ID":"a3bc9902-a999-4ebd-a6d9-23892a188dc5","Type":"ContainerStarted","Data":"4ee7da3f5e5c539579f7aa5399abb8842f941aaafb5584f45f62b890aa7b49b6"} Oct 01 10:14:05 crc kubenswrapper[4837]: I1001 10:14:05.139406 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sk5cn" event={"ID":"a3bc9902-a999-4ebd-a6d9-23892a188dc5","Type":"ContainerStarted","Data":"215c99815631f5dcb13b0cd7e1eea379d093ed6f870bd4d4318ab794e6c58859"} Oct 01 10:14:13 crc kubenswrapper[4837]: I1001 10:14:13.224998 4837 generic.go:334] "Generic (PLEG): container finished" podID="a3bc9902-a999-4ebd-a6d9-23892a188dc5" containerID="4ee7da3f5e5c539579f7aa5399abb8842f941aaafb5584f45f62b890aa7b49b6" exitCode=0 Oct 01 10:14:13 crc kubenswrapper[4837]: I1001 10:14:13.225080 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-sk5cn" event={"ID":"a3bc9902-a999-4ebd-a6d9-23892a188dc5","Type":"ContainerDied","Data":"4ee7da3f5e5c539579f7aa5399abb8842f941aaafb5584f45f62b890aa7b49b6"} Oct 01 10:14:14 crc kubenswrapper[4837]: I1001 10:14:14.818721 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:14:14 crc kubenswrapper[4837]: E1001 10:14:14.819819 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.566169 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635380 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635461 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq2ql\" (UniqueName: \"kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635565 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635641 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635664 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.635751 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf\") pod \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\" (UID: \"a3bc9902-a999-4ebd-a6d9-23892a188dc5\") " Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.637899 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-sk5cn"] Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.638252 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.639202 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.649231 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql" (OuterVolumeSpecName: "kube-api-access-mq2ql") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "kube-api-access-mq2ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.654583 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-sk5cn"] Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.680129 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts" (OuterVolumeSpecName: "scripts") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.683019 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.690807 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.701156 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3bc9902-a999-4ebd-a6d9-23892a188dc5" (UID: "a3bc9902-a999-4ebd-a6d9-23892a188dc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738159 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738212 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a3bc9902-a999-4ebd-a6d9-23892a188dc5-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738225 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738236 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq2ql\" (UniqueName: \"kubernetes.io/projected/a3bc9902-a999-4ebd-a6d9-23892a188dc5-kube-api-access-mq2ql\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738247 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738255 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a3bc9902-a999-4ebd-a6d9-23892a188dc5-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.738266 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3bc9902-a999-4ebd-a6d9-23892a188dc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:14:15 crc kubenswrapper[4837]: I1001 10:14:15.832005 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3bc9902-a999-4ebd-a6d9-23892a188dc5" path="/var/lib/kubelet/pods/a3bc9902-a999-4ebd-a6d9-23892a188dc5/volumes" Oct 01 10:14:16 crc kubenswrapper[4837]: I1001 10:14:16.271569 4837 scope.go:117] "RemoveContainer" containerID="4ee7da3f5e5c539579f7aa5399abb8842f941aaafb5584f45f62b890aa7b49b6" Oct 01 10:14:16 crc kubenswrapper[4837]: I1001 10:14:16.271650 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-sk5cn" Oct 01 10:14:18 crc kubenswrapper[4837]: I1001 10:14:18.021010 4837 scope.go:117] "RemoveContainer" containerID="d5e780133c45c124f08390383be46362f9e411371e9f841a8f986d6cb9ee6b0d" Oct 01 10:14:29 crc kubenswrapper[4837]: I1001 10:14:29.816435 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:14:29 crc kubenswrapper[4837]: E1001 10:14:29.817326 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:14:41 crc kubenswrapper[4837]: I1001 10:14:41.816056 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:14:41 crc kubenswrapper[4837]: E1001 10:14:41.816822 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:14:53 crc kubenswrapper[4837]: I1001 10:14:53.831043 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:14:53 crc kubenswrapper[4837]: E1001 10:14:53.834383 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.176023 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq"] Oct 01 10:15:00 crc kubenswrapper[4837]: E1001 10:15:00.178385 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bc9902-a999-4ebd-a6d9-23892a188dc5" containerName="swift-ring-rebalance" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.178517 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bc9902-a999-4ebd-a6d9-23892a188dc5" containerName="swift-ring-rebalance" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.178952 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3bc9902-a999-4ebd-a6d9-23892a188dc5" containerName="swift-ring-rebalance" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.179983 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.186960 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.200802 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.210742 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq"] Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.235262 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.235310 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vpxg\" (UniqueName: \"kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.235469 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.339077 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.339777 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vpxg\" (UniqueName: \"kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.340286 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.341137 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.346073 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.356422 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vpxg\" (UniqueName: \"kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg\") pod \"collect-profiles-29321895-nszcq\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:00 crc kubenswrapper[4837]: I1001 10:15:00.526010 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:01 crc kubenswrapper[4837]: I1001 10:15:01.266054 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq"] Oct 01 10:15:01 crc kubenswrapper[4837]: I1001 10:15:01.855605 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" event={"ID":"2c2116ee-4786-4dc0-9648-508cc5879dd2","Type":"ContainerStarted","Data":"7568d9eed1072f9a0bcf9e5ef51b53f9f6d6aa4e42e216ef1b834e65f519074e"} Oct 01 10:15:01 crc kubenswrapper[4837]: I1001 10:15:01.855669 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" event={"ID":"2c2116ee-4786-4dc0-9648-508cc5879dd2","Type":"ContainerStarted","Data":"a1c0a01749b123d2a95faddd62dd62e94e0d252ca40d09bf2e71719c68976f62"} Oct 01 10:15:01 crc kubenswrapper[4837]: I1001 10:15:01.878052 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" podStartSLOduration=1.8780352169999999 podStartE2EDuration="1.878035217s" podCreationTimestamp="2025-10-01 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:15:01.875762291 +0000 UTC m=+11358.717369756" watchObservedRunningTime="2025-10-01 10:15:01.878035217 +0000 UTC m=+11358.719642682" Oct 01 10:15:02 crc kubenswrapper[4837]: I1001 10:15:02.869857 4837 generic.go:334] "Generic (PLEG): container finished" podID="2c2116ee-4786-4dc0-9648-508cc5879dd2" containerID="7568d9eed1072f9a0bcf9e5ef51b53f9f6d6aa4e42e216ef1b834e65f519074e" exitCode=0 Oct 01 10:15:02 crc kubenswrapper[4837]: I1001 10:15:02.869930 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" event={"ID":"2c2116ee-4786-4dc0-9648-508cc5879dd2","Type":"ContainerDied","Data":"7568d9eed1072f9a0bcf9e5ef51b53f9f6d6aa4e42e216ef1b834e65f519074e"} Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.373637 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.458751 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vpxg\" (UniqueName: \"kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg\") pod \"2c2116ee-4786-4dc0-9648-508cc5879dd2\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.458867 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume\") pod \"2c2116ee-4786-4dc0-9648-508cc5879dd2\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.458945 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume\") pod \"2c2116ee-4786-4dc0-9648-508cc5879dd2\" (UID: \"2c2116ee-4786-4dc0-9648-508cc5879dd2\") " Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.460051 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c2116ee-4786-4dc0-9648-508cc5879dd2" (UID: "2c2116ee-4786-4dc0-9648-508cc5879dd2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.467998 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c2116ee-4786-4dc0-9648-508cc5879dd2" (UID: "2c2116ee-4786-4dc0-9648-508cc5879dd2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.468082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg" (OuterVolumeSpecName: "kube-api-access-7vpxg") pod "2c2116ee-4786-4dc0-9648-508cc5879dd2" (UID: "2c2116ee-4786-4dc0-9648-508cc5879dd2"). InnerVolumeSpecName "kube-api-access-7vpxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.561113 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2116ee-4786-4dc0-9648-508cc5879dd2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.561150 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2116ee-4786-4dc0-9648-508cc5879dd2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.561159 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vpxg\" (UniqueName: \"kubernetes.io/projected/2c2116ee-4786-4dc0-9648-508cc5879dd2-kube-api-access-7vpxg\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.905897 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" event={"ID":"2c2116ee-4786-4dc0-9648-508cc5879dd2","Type":"ContainerDied","Data":"a1c0a01749b123d2a95faddd62dd62e94e0d252ca40d09bf2e71719c68976f62"} Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.905949 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1c0a01749b123d2a95faddd62dd62e94e0d252ca40d09bf2e71719c68976f62" Oct 01 10:15:05 crc kubenswrapper[4837]: I1001 10:15:05.905964 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-nszcq" Oct 01 10:15:06 crc kubenswrapper[4837]: I1001 10:15:06.463446 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl"] Oct 01 10:15:06 crc kubenswrapper[4837]: I1001 10:15:06.479438 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-drzrl"] Oct 01 10:15:07 crc kubenswrapper[4837]: I1001 10:15:07.831269 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f79b0eb-b8db-4fda-8573-329771458ba8" path="/var/lib/kubelet/pods/1f79b0eb-b8db-4fda-8573-329771458ba8/volumes" Oct 01 10:15:08 crc kubenswrapper[4837]: I1001 10:15:08.816768 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:15:08 crc kubenswrapper[4837]: E1001 10:15:08.817102 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.776533 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-lkpbl"] Oct 01 10:15:15 crc kubenswrapper[4837]: E1001 10:15:15.777535 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2116ee-4786-4dc0-9648-508cc5879dd2" containerName="collect-profiles" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.777549 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2116ee-4786-4dc0-9648-508cc5879dd2" containerName="collect-profiles" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.777795 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2116ee-4786-4dc0-9648-508cc5879dd2" containerName="collect-profiles" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.778504 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.786578 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.786607 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.792678 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-lkpbl"] Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901301 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901380 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901408 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901467 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901495 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7h9z\" (UniqueName: \"kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901543 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:15 crc kubenswrapper[4837]: I1001 10:15:15.901587 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.002956 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003244 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003279 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003343 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003360 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7h9z\" (UniqueName: \"kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003367 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003402 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.003452 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.004156 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.004834 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.009344 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.012481 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.039091 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.040286 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7h9z\" (UniqueName: \"kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z\") pod \"swift-ring-rebalance-debug-lkpbl\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.101274 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:16 crc kubenswrapper[4837]: I1001 10:15:16.914168 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-lkpbl"] Oct 01 10:15:17 crc kubenswrapper[4837]: I1001 10:15:17.020963 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-lkpbl" event={"ID":"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb","Type":"ContainerStarted","Data":"a0cd7b50780fb1ec66e3382e5bcc623f66d7fb8e0786fe71cce11bf886aff32b"} Oct 01 10:15:18 crc kubenswrapper[4837]: I1001 10:15:18.033428 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-lkpbl" event={"ID":"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb","Type":"ContainerStarted","Data":"93ddbbcd19f7432b0c903d9b54edee2c06d3ebd59852693b52ef030c1ff0eda9"} Oct 01 10:15:18 crc kubenswrapper[4837]: I1001 10:15:18.052991 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-lkpbl" podStartSLOduration=3.052976493 podStartE2EDuration="3.052976493s" podCreationTimestamp="2025-10-01 10:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:15:18.046898363 +0000 UTC m=+11374.888505818" watchObservedRunningTime="2025-10-01 10:15:18.052976493 +0000 UTC m=+11374.894583948" Oct 01 10:15:18 crc kubenswrapper[4837]: I1001 10:15:18.118335 4837 scope.go:117] "RemoveContainer" containerID="79a195d572e184aa02b0001aa13c210c0da0627cc5affeb645259c93dadd62ac" Oct 01 10:15:18 crc kubenswrapper[4837]: I1001 10:15:18.144990 4837 scope.go:117] "RemoveContainer" containerID="9423bfcadcba48b768e282dcbe1347a541f5870e591724c2eb37214c9c677123" Oct 01 10:15:23 crc kubenswrapper[4837]: I1001 10:15:23.827015 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:15:23 crc kubenswrapper[4837]: E1001 10:15:23.827647 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.758522 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-4p46l"] Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.760530 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.772331 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-4p46l"] Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873142 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873216 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873293 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873316 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873362 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8xsr\" (UniqueName: \"kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873407 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.873461 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975004 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975176 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975216 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975278 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975293 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975325 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8xsr\" (UniqueName: \"kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975359 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.975851 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.976404 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.976521 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.980944 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.983796 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:26 crc kubenswrapper[4837]: I1001 10:15:26.994331 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:27 crc kubenswrapper[4837]: I1001 10:15:27.004372 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8xsr\" (UniqueName: \"kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr\") pod \"swift-ring-rebalance-debug-4p46l\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:27 crc kubenswrapper[4837]: I1001 10:15:27.119991 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:27 crc kubenswrapper[4837]: I1001 10:15:27.123655 4837 generic.go:334] "Generic (PLEG): container finished" podID="5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" containerID="93ddbbcd19f7432b0c903d9b54edee2c06d3ebd59852693b52ef030c1ff0eda9" exitCode=0 Oct 01 10:15:27 crc kubenswrapper[4837]: I1001 10:15:27.123707 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-lkpbl" event={"ID":"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb","Type":"ContainerDied","Data":"93ddbbcd19f7432b0c903d9b54edee2c06d3ebd59852693b52ef030c1ff0eda9"} Oct 01 10:15:28 crc kubenswrapper[4837]: I1001 10:15:28.065977 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-4p46l"] Oct 01 10:15:28 crc kubenswrapper[4837]: I1001 10:15:28.135851 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4p46l" event={"ID":"916da493-83ed-4efc-b84e-3c91e414218a","Type":"ContainerStarted","Data":"32e9f3e894eefc000e4c778c2ce809d2abbadbfc067ec619584afb41c53ac498"} Oct 01 10:15:29 crc kubenswrapper[4837]: I1001 10:15:29.151349 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4p46l" event={"ID":"916da493-83ed-4efc-b84e-3c91e414218a","Type":"ContainerStarted","Data":"96ac6a2aee16a0b3a3e31b79f354a25e57522dc37ec5c63d1b2d182d6f9a1263"} Oct 01 10:15:29 crc kubenswrapper[4837]: I1001 10:15:29.182028 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-4p46l" podStartSLOduration=3.182003061 podStartE2EDuration="3.182003061s" podCreationTimestamp="2025-10-01 10:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:15:29.172054855 +0000 UTC m=+11386.013662330" watchObservedRunningTime="2025-10-01 10:15:29.182003061 +0000 UTC m=+11386.023610536" Oct 01 10:15:29 crc kubenswrapper[4837]: I1001 10:15:29.998680 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154265 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154451 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154526 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154579 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7h9z\" (UniqueName: \"kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154618 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154665 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts\") pod \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\" (UID: \"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb\") " Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.154974 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.155325 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.161743 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.161790 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z" (OuterVolumeSpecName: "kube-api-access-l7h9z") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "kube-api-access-l7h9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.185104 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-lkpbl" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.185179 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-lkpbl" event={"ID":"5d2b6854-fb7d-4a91-ace2-1c25c131f9fb","Type":"ContainerDied","Data":"a0cd7b50780fb1ec66e3382e5bcc623f66d7fb8e0786fe71cce11bf886aff32b"} Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.185326 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0cd7b50780fb1ec66e3382e5bcc623f66d7fb8e0786fe71cce11bf886aff32b" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.194019 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.201449 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.203978 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts" (OuterVolumeSpecName: "scripts") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.204552 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" (UID: "5d2b6854-fb7d-4a91-ace2-1c25c131f9fb"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257733 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257755 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257767 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7h9z\" (UniqueName: \"kubernetes.io/projected/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-kube-api-access-l7h9z\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257775 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257783 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:30 crc kubenswrapper[4837]: I1001 10:15:30.257792 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d2b6854-fb7d-4a91-ace2-1c25c131f9fb-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:34 crc kubenswrapper[4837]: I1001 10:15:34.237426 4837 generic.go:334] "Generic (PLEG): container finished" podID="916da493-83ed-4efc-b84e-3c91e414218a" containerID="96ac6a2aee16a0b3a3e31b79f354a25e57522dc37ec5c63d1b2d182d6f9a1263" exitCode=0 Oct 01 10:15:34 crc kubenswrapper[4837]: I1001 10:15:34.237502 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4p46l" event={"ID":"916da493-83ed-4efc-b84e-3c91e414218a","Type":"ContainerDied","Data":"96ac6a2aee16a0b3a3e31b79f354a25e57522dc37ec5c63d1b2d182d6f9a1263"} Oct 01 10:15:34 crc kubenswrapper[4837]: I1001 10:15:34.817013 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:15:34 crc kubenswrapper[4837]: E1001 10:15:34.817370 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:15:36 crc kubenswrapper[4837]: I1001 10:15:36.971482 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.028849 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-4p46l"] Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.038271 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-4p46l"] Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.115562 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.115727 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.115878 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.115945 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.115987 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.116059 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8xsr\" (UniqueName: \"kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.116086 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle\") pod \"916da493-83ed-4efc-b84e-3c91e414218a\" (UID: \"916da493-83ed-4efc-b84e-3c91e414218a\") " Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.116866 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.118781 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.124285 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr" (OuterVolumeSpecName: "kube-api-access-k8xsr") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "kube-api-access-k8xsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.148379 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.148878 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.159003 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts" (OuterVolumeSpecName: "scripts") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.161404 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "916da493-83ed-4efc-b84e-3c91e414218a" (UID: "916da493-83ed-4efc-b84e-3c91e414218a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219389 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219432 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219446 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219457 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/916da493-83ed-4efc-b84e-3c91e414218a-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219468 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8xsr\" (UniqueName: \"kubernetes.io/projected/916da493-83ed-4efc-b84e-3c91e414218a-kube-api-access-k8xsr\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219481 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/916da493-83ed-4efc-b84e-3c91e414218a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.219491 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/916da493-83ed-4efc-b84e-3c91e414218a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.271526 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32e9f3e894eefc000e4c778c2ce809d2abbadbfc067ec619584afb41c53ac498" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.271609 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4p46l" Oct 01 10:15:37 crc kubenswrapper[4837]: I1001 10:15:37.833247 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="916da493-83ed-4efc-b84e-3c91e414218a" path="/var/lib/kubelet/pods/916da493-83ed-4efc-b84e-3c91e414218a/volumes" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.995467 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-cvsn8"] Oct 01 10:15:40 crc kubenswrapper[4837]: E1001 10:15:40.996905 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.996928 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: E1001 10:15:40.996958 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916da493-83ed-4efc-b84e-3c91e414218a" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.996969 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="916da493-83ed-4efc-b84e-3c91e414218a" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.997259 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2b6854-fb7d-4a91-ace2-1c25c131f9fb" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.997282 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="916da493-83ed-4efc-b84e-3c91e414218a" containerName="swift-ring-rebalance" Oct 01 10:15:40 crc kubenswrapper[4837]: I1001 10:15:40.998308 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.000283 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.000763 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.014310 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-cvsn8"] Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113423 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfxww\" (UniqueName: \"kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113533 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113581 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113645 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113670 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113807 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.113911 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215643 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215753 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfxww\" (UniqueName: \"kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215809 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215842 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215890 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215909 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.215957 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.216286 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.216918 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.217494 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.222285 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.224053 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.225019 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.235398 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfxww\" (UniqueName: \"kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww\") pod \"swift-ring-rebalance-debug-cvsn8\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:41 crc kubenswrapper[4837]: I1001 10:15:41.335338 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:42 crc kubenswrapper[4837]: I1001 10:15:42.036295 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-cvsn8"] Oct 01 10:15:42 crc kubenswrapper[4837]: I1001 10:15:42.409034 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-cvsn8" event={"ID":"e37424e9-aab0-4822-aa2b-fb007a574771","Type":"ContainerStarted","Data":"3906f614001dd109df93dba78087c0c78f6c0aa41c3dadabcd220e8ea90a1504"} Oct 01 10:15:42 crc kubenswrapper[4837]: I1001 10:15:42.409414 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-cvsn8" event={"ID":"e37424e9-aab0-4822-aa2b-fb007a574771","Type":"ContainerStarted","Data":"34a0b2abd75f5cc6e2b2a23f8ab78f32e3f31df2b3c267ded82a8df990f586fd"} Oct 01 10:15:42 crc kubenswrapper[4837]: I1001 10:15:42.431567 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-cvsn8" podStartSLOduration=2.431548941 podStartE2EDuration="2.431548941s" podCreationTimestamp="2025-10-01 10:15:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:15:42.428634549 +0000 UTC m=+11399.270242004" watchObservedRunningTime="2025-10-01 10:15:42.431548941 +0000 UTC m=+11399.273156396" Oct 01 10:15:48 crc kubenswrapper[4837]: I1001 10:15:48.816757 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:15:48 crc kubenswrapper[4837]: E1001 10:15:48.817407 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:15:52 crc kubenswrapper[4837]: I1001 10:15:52.542439 4837 generic.go:334] "Generic (PLEG): container finished" podID="e37424e9-aab0-4822-aa2b-fb007a574771" containerID="3906f614001dd109df93dba78087c0c78f6c0aa41c3dadabcd220e8ea90a1504" exitCode=0 Oct 01 10:15:52 crc kubenswrapper[4837]: I1001 10:15:52.542840 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-cvsn8" event={"ID":"e37424e9-aab0-4822-aa2b-fb007a574771","Type":"ContainerDied","Data":"3906f614001dd109df93dba78087c0c78f6c0aa41c3dadabcd220e8ea90a1504"} Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.359294 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.404479 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-cvsn8"] Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.415065 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-cvsn8"] Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.533457 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.533538 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.533626 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfxww\" (UniqueName: \"kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.533671 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.533724 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.534365 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.534546 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.534803 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.534917 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf\") pod \"e37424e9-aab0-4822-aa2b-fb007a574771\" (UID: \"e37424e9-aab0-4822-aa2b-fb007a574771\") " Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.535765 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e37424e9-aab0-4822-aa2b-fb007a574771-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.535789 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.539808 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww" (OuterVolumeSpecName: "kube-api-access-cfxww") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "kube-api-access-cfxww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.573499 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.580547 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34a0b2abd75f5cc6e2b2a23f8ab78f32e3f31df2b3c267ded82a8df990f586fd" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.580730 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-cvsn8" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.589821 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.589955 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.590743 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts" (OuterVolumeSpecName: "scripts") pod "e37424e9-aab0-4822-aa2b-fb007a574771" (UID: "e37424e9-aab0-4822-aa2b-fb007a574771"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.639523 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfxww\" (UniqueName: \"kubernetes.io/projected/e37424e9-aab0-4822-aa2b-fb007a574771-kube-api-access-cfxww\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.639561 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.639571 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e37424e9-aab0-4822-aa2b-fb007a574771-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.639581 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.639591 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37424e9-aab0-4822-aa2b-fb007a574771-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:55 crc kubenswrapper[4837]: I1001 10:15:55.831014 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37424e9-aab0-4822-aa2b-fb007a574771" path="/var/lib/kubelet/pods/e37424e9-aab0-4822-aa2b-fb007a574771/volumes" Oct 01 10:16:00 crc kubenswrapper[4837]: I1001 10:16:00.816449 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:16:00 crc kubenswrapper[4837]: E1001 10:16:00.817405 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:16:12 crc kubenswrapper[4837]: I1001 10:16:12.816078 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:16:12 crc kubenswrapper[4837]: E1001 10:16:12.816905 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:16:26 crc kubenswrapper[4837]: I1001 10:16:26.817154 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:16:26 crc kubenswrapper[4837]: E1001 10:16:26.818173 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:16:40 crc kubenswrapper[4837]: I1001 10:16:40.816042 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:16:40 crc kubenswrapper[4837]: E1001 10:16:40.816872 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:16:51 crc kubenswrapper[4837]: I1001 10:16:51.816958 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:16:51 crc kubenswrapper[4837]: E1001 10:16:51.817812 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.632727 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-xxn92"] Oct 01 10:16:55 crc kubenswrapper[4837]: E1001 10:16:55.635237 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37424e9-aab0-4822-aa2b-fb007a574771" containerName="swift-ring-rebalance" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.635262 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37424e9-aab0-4822-aa2b-fb007a574771" containerName="swift-ring-rebalance" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.635573 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37424e9-aab0-4822-aa2b-fb007a574771" containerName="swift-ring-rebalance" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.636731 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.641643 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.641682 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660322 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660394 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660450 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660490 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660564 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcgmp\" (UniqueName: \"kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660675 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.660722 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.663815 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-xxn92"] Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762475 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762567 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762597 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762636 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762661 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762686 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcgmp\" (UniqueName: \"kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.762784 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.763559 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.763845 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.763897 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.773363 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.775201 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.777718 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.783240 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcgmp\" (UniqueName: \"kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp\") pod \"swift-ring-rebalance-debug-xxn92\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:55 crc kubenswrapper[4837]: I1001 10:16:55.965023 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:16:56 crc kubenswrapper[4837]: I1001 10:16:56.878085 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-xxn92"] Oct 01 10:16:57 crc kubenswrapper[4837]: I1001 10:16:57.403673 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xxn92" event={"ID":"4e5e4ce8-3094-4a98-b27f-e0e597deab7a","Type":"ContainerStarted","Data":"ec5028f8f737edf2d9958b3c2c288c4bddbc88d69ca6de0230ce56a186fa85bf"} Oct 01 10:16:57 crc kubenswrapper[4837]: I1001 10:16:57.404064 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xxn92" event={"ID":"4e5e4ce8-3094-4a98-b27f-e0e597deab7a","Type":"ContainerStarted","Data":"5c7fc2c4ad1b26f61b14d108f8fd86aff2179112eeb72bba9640c9ce1fe46b58"} Oct 01 10:16:57 crc kubenswrapper[4837]: I1001 10:16:57.429537 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-xxn92" podStartSLOduration=2.42951811 podStartE2EDuration="2.42951811s" podCreationTimestamp="2025-10-01 10:16:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:16:57.425362468 +0000 UTC m=+11474.266969943" watchObservedRunningTime="2025-10-01 10:16:57.42951811 +0000 UTC m=+11474.271125565" Oct 01 10:17:06 crc kubenswrapper[4837]: I1001 10:17:06.816700 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:17:07 crc kubenswrapper[4837]: I1001 10:17:07.514879 4837 generic.go:334] "Generic (PLEG): container finished" podID="4e5e4ce8-3094-4a98-b27f-e0e597deab7a" containerID="ec5028f8f737edf2d9958b3c2c288c4bddbc88d69ca6de0230ce56a186fa85bf" exitCode=0 Oct 01 10:17:07 crc kubenswrapper[4837]: I1001 10:17:07.515121 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-xxn92" event={"ID":"4e5e4ce8-3094-4a98-b27f-e0e597deab7a","Type":"ContainerDied","Data":"ec5028f8f737edf2d9958b3c2c288c4bddbc88d69ca6de0230ce56a186fa85bf"} Oct 01 10:17:07 crc kubenswrapper[4837]: I1001 10:17:07.517577 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995"} Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.380258 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.441414 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-xxn92"] Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.453979 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-xxn92"] Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.502745 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.502809 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.503606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcgmp\" (UniqueName: \"kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.503733 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.503843 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.503892 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.503930 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices\") pod \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\" (UID: \"4e5e4ce8-3094-4a98-b27f-e0e597deab7a\") " Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.504653 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.504993 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.509938 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp" (OuterVolumeSpecName: "kube-api-access-jcgmp") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "kube-api-access-jcgmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.535996 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts" (OuterVolumeSpecName: "scripts") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.545047 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.545187 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.557009 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e5e4ce8-3094-4a98-b27f-e0e597deab7a" (UID: "4e5e4ce8-3094-4a98-b27f-e0e597deab7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.564476 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c7fc2c4ad1b26f61b14d108f8fd86aff2179112eeb72bba9640c9ce1fe46b58" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.564566 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-xxn92" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612422 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612452 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612462 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612470 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612478 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612488 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcgmp\" (UniqueName: \"kubernetes.io/projected/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-kube-api-access-jcgmp\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:10 crc kubenswrapper[4837]: I1001 10:17:10.612497 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4e5e4ce8-3094-4a98-b27f-e0e597deab7a-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:17:11 crc kubenswrapper[4837]: I1001 10:17:11.827093 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e5e4ce8-3094-4a98-b27f-e0e597deab7a" path="/var/lib/kubelet/pods/4e5e4ce8-3094-4a98-b27f-e0e597deab7a/volumes" Oct 01 10:17:12 crc kubenswrapper[4837]: I1001 10:17:12.802107 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-bd98cb97c-4l4pg" podUID="c629d2d6-9a91-4a51-9504-1524ee0bc8ed" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.352563 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:00 crc kubenswrapper[4837]: E1001 10:18:00.354487 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e5e4ce8-3094-4a98-b27f-e0e597deab7a" containerName="swift-ring-rebalance" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.354514 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e5e4ce8-3094-4a98-b27f-e0e597deab7a" containerName="swift-ring-rebalance" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.355302 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e5e4ce8-3094-4a98-b27f-e0e597deab7a" containerName="swift-ring-rebalance" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.358024 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.382340 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.491260 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.491352 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djlgk\" (UniqueName: \"kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.491504 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.594935 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.595030 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.595458 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.595550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djlgk\" (UniqueName: \"kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.595652 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.620197 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djlgk\" (UniqueName: \"kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk\") pod \"community-operators-jgx74\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:00 crc kubenswrapper[4837]: I1001 10:18:00.685737 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:01 crc kubenswrapper[4837]: W1001 10:18:01.462176 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5b09395_aacb_450f_b58b_0295651d4049.slice/crio-11a5f7207ac5bfa3b97644c54477c63d744713ee9c8e32537f1d995bac2df40e WatchSource:0}: Error finding container 11a5f7207ac5bfa3b97644c54477c63d744713ee9c8e32537f1d995bac2df40e: Status 404 returned error can't find the container with id 11a5f7207ac5bfa3b97644c54477c63d744713ee9c8e32537f1d995bac2df40e Oct 01 10:18:01 crc kubenswrapper[4837]: I1001 10:18:01.464392 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:02 crc kubenswrapper[4837]: I1001 10:18:02.166254 4837 generic.go:334] "Generic (PLEG): container finished" podID="f5b09395-aacb-450f-b58b-0295651d4049" containerID="195889c6c13571f47de65b3a420819881f070457304d37823bf9e11930fbd434" exitCode=0 Oct 01 10:18:02 crc kubenswrapper[4837]: I1001 10:18:02.166347 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerDied","Data":"195889c6c13571f47de65b3a420819881f070457304d37823bf9e11930fbd434"} Oct 01 10:18:02 crc kubenswrapper[4837]: I1001 10:18:02.166628 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerStarted","Data":"11a5f7207ac5bfa3b97644c54477c63d744713ee9c8e32537f1d995bac2df40e"} Oct 01 10:18:02 crc kubenswrapper[4837]: I1001 10:18:02.168364 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:18:03 crc kubenswrapper[4837]: I1001 10:18:03.188546 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerStarted","Data":"6efe5dc90e6354b52383b4a0b083a400baf0d2b3dace52f618cc8483d2d885d5"} Oct 01 10:18:07 crc kubenswrapper[4837]: I1001 10:18:07.243237 4837 generic.go:334] "Generic (PLEG): container finished" podID="f5b09395-aacb-450f-b58b-0295651d4049" containerID="6efe5dc90e6354b52383b4a0b083a400baf0d2b3dace52f618cc8483d2d885d5" exitCode=0 Oct 01 10:18:07 crc kubenswrapper[4837]: I1001 10:18:07.243339 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerDied","Data":"6efe5dc90e6354b52383b4a0b083a400baf0d2b3dace52f618cc8483d2d885d5"} Oct 01 10:18:08 crc kubenswrapper[4837]: I1001 10:18:08.275026 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerStarted","Data":"df731298ae2345ff8c871f260e39c72d1a8d6a57730ade20192dee04919a9e5e"} Oct 01 10:18:08 crc kubenswrapper[4837]: I1001 10:18:08.309208 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jgx74" podStartSLOduration=2.561164404 podStartE2EDuration="8.309187606s" podCreationTimestamp="2025-10-01 10:18:00 +0000 UTC" firstStartedPulling="2025-10-01 10:18:02.167954576 +0000 UTC m=+11539.009562071" lastFinishedPulling="2025-10-01 10:18:07.915977808 +0000 UTC m=+11544.757585273" observedRunningTime="2025-10-01 10:18:08.305735921 +0000 UTC m=+11545.147343386" watchObservedRunningTime="2025-10-01 10:18:08.309187606 +0000 UTC m=+11545.150795081" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.616021 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-p7vsb"] Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.619230 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.623606 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.624114 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.644853 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-p7vsb"] Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.647833 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkw8v\" (UniqueName: \"kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.647883 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.647980 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.648083 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.648112 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.648183 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.648209 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.686732 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.686789 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752075 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752147 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752173 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkw8v\" (UniqueName: \"kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752196 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752253 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752311 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.752339 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.756838 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.757470 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.761409 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.767715 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.774070 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.774147 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.778391 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkw8v\" (UniqueName: \"kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v\") pod \"swift-ring-rebalance-debug-p7vsb\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:10 crc kubenswrapper[4837]: I1001 10:18:10.940264 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:11 crc kubenswrapper[4837]: I1001 10:18:11.667341 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-p7vsb"] Oct 01 10:18:11 crc kubenswrapper[4837]: I1001 10:18:11.743349 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jgx74" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" probeResult="failure" output=< Oct 01 10:18:11 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:18:11 crc kubenswrapper[4837]: > Oct 01 10:18:12 crc kubenswrapper[4837]: I1001 10:18:12.320279 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-p7vsb" event={"ID":"041200f3-6d68-46a9-8651-931da335026b","Type":"ContainerStarted","Data":"201acb359643a6e002344bc1c84223c2e322389d2928e530c1630b61e30828aa"} Oct 01 10:18:12 crc kubenswrapper[4837]: I1001 10:18:12.320623 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-p7vsb" event={"ID":"041200f3-6d68-46a9-8651-931da335026b","Type":"ContainerStarted","Data":"b63be327d4a229553b7bb83933cf400059c4cfa41e5766d8d78e01615151d534"} Oct 01 10:18:12 crc kubenswrapper[4837]: I1001 10:18:12.342224 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-p7vsb" podStartSLOduration=2.342205751 podStartE2EDuration="2.342205751s" podCreationTimestamp="2025-10-01 10:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:18:12.333542007 +0000 UTC m=+11549.175149472" watchObservedRunningTime="2025-10-01 10:18:12.342205751 +0000 UTC m=+11549.183813216" Oct 01 10:18:18 crc kubenswrapper[4837]: I1001 10:18:18.307015 4837 scope.go:117] "RemoveContainer" containerID="637b41e73303dc1610abaed84f5f841bf46ca6467531cd6b27b741e7e7b6efe5" Oct 01 10:18:21 crc kubenswrapper[4837]: I1001 10:18:21.741133 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jgx74" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" probeResult="failure" output=< Oct 01 10:18:21 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:18:21 crc kubenswrapper[4837]: > Oct 01 10:18:23 crc kubenswrapper[4837]: I1001 10:18:23.445330 4837 generic.go:334] "Generic (PLEG): container finished" podID="041200f3-6d68-46a9-8651-931da335026b" containerID="201acb359643a6e002344bc1c84223c2e322389d2928e530c1630b61e30828aa" exitCode=0 Oct 01 10:18:23 crc kubenswrapper[4837]: I1001 10:18:23.445403 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-p7vsb" event={"ID":"041200f3-6d68-46a9-8651-931da335026b","Type":"ContainerDied","Data":"201acb359643a6e002344bc1c84223c2e322389d2928e530c1630b61e30828aa"} Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.327435 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.415412 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-p7vsb"] Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.429575 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-p7vsb"] Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.488343 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b63be327d4a229553b7bb83933cf400059c4cfa41e5766d8d78e01615151d534" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.488395 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-p7vsb" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.521484 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.521542 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.521620 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.521808 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkw8v\" (UniqueName: \"kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.522048 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.522088 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.522158 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift\") pod \"041200f3-6d68-46a9-8651-931da335026b\" (UID: \"041200f3-6d68-46a9-8651-931da335026b\") " Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.523515 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.523515 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.543614 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v" (OuterVolumeSpecName: "kube-api-access-tkw8v") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "kube-api-access-tkw8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.559168 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.561863 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.570774 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts" (OuterVolumeSpecName: "scripts") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.573397 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "041200f3-6d68-46a9-8651-931da335026b" (UID: "041200f3-6d68-46a9-8651-931da335026b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626263 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626297 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626309 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/041200f3-6d68-46a9-8651-931da335026b-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626321 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkw8v\" (UniqueName: \"kubernetes.io/projected/041200f3-6d68-46a9-8651-931da335026b-kube-api-access-tkw8v\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626333 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626346 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/041200f3-6d68-46a9-8651-931da335026b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.626356 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/041200f3-6d68-46a9-8651-931da335026b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.863580 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-kcdr2"] Oct 01 10:18:26 crc kubenswrapper[4837]: E1001 10:18:26.864008 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041200f3-6d68-46a9-8651-931da335026b" containerName="swift-ring-rebalance" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.864025 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="041200f3-6d68-46a9-8651-931da335026b" containerName="swift-ring-rebalance" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.864265 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="041200f3-6d68-46a9-8651-931da335026b" containerName="swift-ring-rebalance" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.865683 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.868324 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.869377 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.885371 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-kcdr2"] Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.931945 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932093 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932263 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932582 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjrf9\" (UniqueName: \"kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932658 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:26 crc kubenswrapper[4837]: I1001 10:18:26.932785 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034110 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034203 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034314 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjrf9\" (UniqueName: \"kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034356 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034402 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034433 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.034478 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.035273 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.035751 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.035968 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.038632 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.039122 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.041330 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.052466 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjrf9\" (UniqueName: \"kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9\") pod \"swift-ring-rebalance-debug-kcdr2\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.185909 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:27 crc kubenswrapper[4837]: I1001 10:18:27.835166 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041200f3-6d68-46a9-8651-931da335026b" path="/var/lib/kubelet/pods/041200f3-6d68-46a9-8651-931da335026b/volumes" Oct 01 10:18:28 crc kubenswrapper[4837]: I1001 10:18:28.142781 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-kcdr2"] Oct 01 10:18:28 crc kubenswrapper[4837]: W1001 10:18:28.147130 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbed42153_842f_4c78_8604_fac284a72fa3.slice/crio-4f4501251729faeaa2597b575268b0c84d304c368bcb366caeca255b468cfcab WatchSource:0}: Error finding container 4f4501251729faeaa2597b575268b0c84d304c368bcb366caeca255b468cfcab: Status 404 returned error can't find the container with id 4f4501251729faeaa2597b575268b0c84d304c368bcb366caeca255b468cfcab Oct 01 10:18:28 crc kubenswrapper[4837]: I1001 10:18:28.516366 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-kcdr2" event={"ID":"bed42153-842f-4c78-8604-fac284a72fa3","Type":"ContainerStarted","Data":"f2d8273b78ac8fcbf6f44debcdc110989cf1c42bad270d86beee2b94a0986bc4"} Oct 01 10:18:28 crc kubenswrapper[4837]: I1001 10:18:28.516421 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-kcdr2" event={"ID":"bed42153-842f-4c78-8604-fac284a72fa3","Type":"ContainerStarted","Data":"4f4501251729faeaa2597b575268b0c84d304c368bcb366caeca255b468cfcab"} Oct 01 10:18:28 crc kubenswrapper[4837]: I1001 10:18:28.542331 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-kcdr2" podStartSLOduration=2.542313382 podStartE2EDuration="2.542313382s" podCreationTimestamp="2025-10-01 10:18:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:18:28.53090459 +0000 UTC m=+11565.372512045" watchObservedRunningTime="2025-10-01 10:18:28.542313382 +0000 UTC m=+11565.383920837" Oct 01 10:18:30 crc kubenswrapper[4837]: I1001 10:18:30.750385 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:30 crc kubenswrapper[4837]: I1001 10:18:30.817283 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:31 crc kubenswrapper[4837]: I1001 10:18:31.548982 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:32 crc kubenswrapper[4837]: I1001 10:18:32.575119 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jgx74" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" containerID="cri-o://df731298ae2345ff8c871f260e39c72d1a8d6a57730ade20192dee04919a9e5e" gracePeriod=2 Oct 01 10:18:33 crc kubenswrapper[4837]: I1001 10:18:33.601089 4837 generic.go:334] "Generic (PLEG): container finished" podID="bed42153-842f-4c78-8604-fac284a72fa3" containerID="f2d8273b78ac8fcbf6f44debcdc110989cf1c42bad270d86beee2b94a0986bc4" exitCode=0 Oct 01 10:18:33 crc kubenswrapper[4837]: I1001 10:18:33.601322 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-kcdr2" event={"ID":"bed42153-842f-4c78-8604-fac284a72fa3","Type":"ContainerDied","Data":"f2d8273b78ac8fcbf6f44debcdc110989cf1c42bad270d86beee2b94a0986bc4"} Oct 01 10:18:33 crc kubenswrapper[4837]: I1001 10:18:33.607411 4837 generic.go:334] "Generic (PLEG): container finished" podID="f5b09395-aacb-450f-b58b-0295651d4049" containerID="df731298ae2345ff8c871f260e39c72d1a8d6a57730ade20192dee04919a9e5e" exitCode=0 Oct 01 10:18:33 crc kubenswrapper[4837]: I1001 10:18:33.607444 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerDied","Data":"df731298ae2345ff8c871f260e39c72d1a8d6a57730ade20192dee04919a9e5e"} Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.164864 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.309727 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djlgk\" (UniqueName: \"kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk\") pod \"f5b09395-aacb-450f-b58b-0295651d4049\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.309821 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities\") pod \"f5b09395-aacb-450f-b58b-0295651d4049\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.309877 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content\") pod \"f5b09395-aacb-450f-b58b-0295651d4049\" (UID: \"f5b09395-aacb-450f-b58b-0295651d4049\") " Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.310609 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities" (OuterVolumeSpecName: "utilities") pod "f5b09395-aacb-450f-b58b-0295651d4049" (UID: "f5b09395-aacb-450f-b58b-0295651d4049"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.317029 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk" (OuterVolumeSpecName: "kube-api-access-djlgk") pod "f5b09395-aacb-450f-b58b-0295651d4049" (UID: "f5b09395-aacb-450f-b58b-0295651d4049"). InnerVolumeSpecName "kube-api-access-djlgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.355602 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5b09395-aacb-450f-b58b-0295651d4049" (UID: "f5b09395-aacb-450f-b58b-0295651d4049"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.412364 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.412398 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b09395-aacb-450f-b58b-0295651d4049-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.412409 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djlgk\" (UniqueName: \"kubernetes.io/projected/f5b09395-aacb-450f-b58b-0295651d4049-kube-api-access-djlgk\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.625500 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jgx74" event={"ID":"f5b09395-aacb-450f-b58b-0295651d4049","Type":"ContainerDied","Data":"11a5f7207ac5bfa3b97644c54477c63d744713ee9c8e32537f1d995bac2df40e"} Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.625564 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jgx74" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.626094 4837 scope.go:117] "RemoveContainer" containerID="df731298ae2345ff8c871f260e39c72d1a8d6a57730ade20192dee04919a9e5e" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.679273 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.694259 4837 scope.go:117] "RemoveContainer" containerID="6efe5dc90e6354b52383b4a0b083a400baf0d2b3dace52f618cc8483d2d885d5" Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.694775 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jgx74"] Oct 01 10:18:34 crc kubenswrapper[4837]: I1001 10:18:34.752636 4837 scope.go:117] "RemoveContainer" containerID="195889c6c13571f47de65b3a420819881f070457304d37823bf9e11930fbd434" Oct 01 10:18:35 crc kubenswrapper[4837]: I1001 10:18:35.832519 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b09395-aacb-450f-b58b-0295651d4049" path="/var/lib/kubelet/pods/f5b09395-aacb-450f-b58b-0295651d4049/volumes" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.085943 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.133745 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-kcdr2"] Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.151456 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-kcdr2"] Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.254947 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255170 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255214 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255254 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255287 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255348 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjrf9\" (UniqueName: \"kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.255466 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle\") pod \"bed42153-842f-4c78-8604-fac284a72fa3\" (UID: \"bed42153-842f-4c78-8604-fac284a72fa3\") " Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.256306 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.256615 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.269004 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9" (OuterVolumeSpecName: "kube-api-access-bjrf9") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "kube-api-access-bjrf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.286638 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts" (OuterVolumeSpecName: "scripts") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.306625 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.315076 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.324074 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bed42153-842f-4c78-8604-fac284a72fa3" (UID: "bed42153-842f-4c78-8604-fac284a72fa3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357442 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357478 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357489 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357497 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bed42153-842f-4c78-8604-fac284a72fa3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357505 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bed42153-842f-4c78-8604-fac284a72fa3-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357512 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bed42153-842f-4c78-8604-fac284a72fa3-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.357522 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjrf9\" (UniqueName: \"kubernetes.io/projected/bed42153-842f-4c78-8604-fac284a72fa3-kube-api-access-bjrf9\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.656055 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f4501251729faeaa2597b575268b0c84d304c368bcb366caeca255b468cfcab" Oct 01 10:18:36 crc kubenswrapper[4837]: I1001 10:18:36.656150 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-kcdr2" Oct 01 10:18:37 crc kubenswrapper[4837]: I1001 10:18:37.826156 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bed42153-842f-4c78-8604-fac284a72fa3" path="/var/lib/kubelet/pods/bed42153-842f-4c78-8604-fac284a72fa3/volumes" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.985674 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-6pmtw"] Oct 01 10:18:39 crc kubenswrapper[4837]: E1001 10:18:39.986389 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bed42153-842f-4c78-8604-fac284a72fa3" containerName="swift-ring-rebalance" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986404 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="bed42153-842f-4c78-8604-fac284a72fa3" containerName="swift-ring-rebalance" Oct 01 10:18:39 crc kubenswrapper[4837]: E1001 10:18:39.986421 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="extract-utilities" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986427 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="extract-utilities" Oct 01 10:18:39 crc kubenswrapper[4837]: E1001 10:18:39.986466 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986474 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" Oct 01 10:18:39 crc kubenswrapper[4837]: E1001 10:18:39.986493 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="extract-content" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986500 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="extract-content" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986709 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b09395-aacb-450f-b58b-0295651d4049" containerName="registry-server" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.986726 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="bed42153-842f-4c78-8604-fac284a72fa3" containerName="swift-ring-rebalance" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.987465 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.989769 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:18:39 crc kubenswrapper[4837]: I1001 10:18:39.989789 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.001608 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-6pmtw"] Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.132661 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.132746 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.132841 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.132956 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.133189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.133299 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfmj8\" (UniqueName: \"kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.133350 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235466 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235567 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235614 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfmj8\" (UniqueName: \"kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235676 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235806 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.235859 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.236705 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.236435 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.236605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.236475 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.242931 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.247790 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.251313 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.265194 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfmj8\" (UniqueName: \"kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8\") pod \"swift-ring-rebalance-debug-6pmtw\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:40 crc kubenswrapper[4837]: I1001 10:18:40.308206 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:41 crc kubenswrapper[4837]: I1001 10:18:41.052825 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-6pmtw"] Oct 01 10:18:41 crc kubenswrapper[4837]: I1001 10:18:41.717799 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-6pmtw" event={"ID":"b40afda2-a734-4619-a0e1-980d12f72780","Type":"ContainerStarted","Data":"9d0cc08c4d77c4d6a0db5327ed0c4973d0ac27a6104936256e5a658300645b53"} Oct 01 10:18:41 crc kubenswrapper[4837]: I1001 10:18:41.718161 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-6pmtw" event={"ID":"b40afda2-a734-4619-a0e1-980d12f72780","Type":"ContainerStarted","Data":"4fc1ab3b81e8497240956dc1be94139ae72624213d36a710ab604f5dbbd02e77"} Oct 01 10:18:41 crc kubenswrapper[4837]: I1001 10:18:41.735407 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-6pmtw" podStartSLOduration=2.7353920819999997 podStartE2EDuration="2.735392082s" podCreationTimestamp="2025-10-01 10:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:18:41.731911855 +0000 UTC m=+11578.573519330" watchObservedRunningTime="2025-10-01 10:18:41.735392082 +0000 UTC m=+11578.576999537" Oct 01 10:18:53 crc kubenswrapper[4837]: I1001 10:18:53.858122 4837 generic.go:334] "Generic (PLEG): container finished" podID="b40afda2-a734-4619-a0e1-980d12f72780" containerID="9d0cc08c4d77c4d6a0db5327ed0c4973d0ac27a6104936256e5a658300645b53" exitCode=0 Oct 01 10:18:53 crc kubenswrapper[4837]: I1001 10:18:53.858195 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-6pmtw" event={"ID":"b40afda2-a734-4619-a0e1-980d12f72780","Type":"ContainerDied","Data":"9d0cc08c4d77c4d6a0db5327ed0c4973d0ac27a6104936256e5a658300645b53"} Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.787606 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.832982 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-6pmtw"] Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.851222 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-6pmtw"] Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892488 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892591 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892617 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892713 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892777 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfmj8\" (UniqueName: \"kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.892868 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf\") pod \"b40afda2-a734-4619-a0e1-980d12f72780\" (UID: \"b40afda2-a734-4619-a0e1-980d12f72780\") " Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.893074 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.893393 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.894029 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.894528 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fc1ab3b81e8497240956dc1be94139ae72624213d36a710ab604f5dbbd02e77" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.894600 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-6pmtw" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.905017 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8" (OuterVolumeSpecName: "kube-api-access-vfmj8") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "kube-api-access-vfmj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.928831 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.932266 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.948740 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts" (OuterVolumeSpecName: "scripts") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.951788 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b40afda2-a734-4619-a0e1-980d12f72780" (UID: "b40afda2-a734-4619-a0e1-980d12f72780"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995183 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b40afda2-a734-4619-a0e1-980d12f72780-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995832 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995870 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995904 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b40afda2-a734-4619-a0e1-980d12f72780-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995927 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfmj8\" (UniqueName: \"kubernetes.io/projected/b40afda2-a734-4619-a0e1-980d12f72780-kube-api-access-vfmj8\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:56 crc kubenswrapper[4837]: I1001 10:18:56.995939 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b40afda2-a734-4619-a0e1-980d12f72780-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:57 crc kubenswrapper[4837]: I1001 10:18:57.826905 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40afda2-a734-4619-a0e1-980d12f72780" path="/var/lib/kubelet/pods/b40afda2-a734-4619-a0e1-980d12f72780/volumes" Oct 01 10:19:18 crc kubenswrapper[4837]: I1001 10:19:18.388510 4837 scope.go:117] "RemoveContainer" containerID="482d0a8fe52679d2c89dbdb4cbef1b60bfee2b4b19d8276bebb72ceb7be21e08" Oct 01 10:19:23 crc kubenswrapper[4837]: I1001 10:19:23.078976 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:19:23 crc kubenswrapper[4837]: I1001 10:19:23.079551 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:19:53 crc kubenswrapper[4837]: I1001 10:19:53.079454 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:19:53 crc kubenswrapper[4837]: I1001 10:19:53.079978 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.079640 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.080101 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.080146 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.080914 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.080957 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995" gracePeriod=600 Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.933665 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995" exitCode=0 Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.935178 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995"} Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.935232 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce"} Oct 01 10:20:23 crc kubenswrapper[4837]: I1001 10:20:23.935260 4837 scope.go:117] "RemoveContainer" containerID="8744ae10d05c366b7135ad9d026f5981254654f41d0085ac940562f388beba9a" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.881289 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-g5758"] Oct 01 10:22:03 crc kubenswrapper[4837]: E1001 10:22:03.883301 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40afda2-a734-4619-a0e1-980d12f72780" containerName="swift-ring-rebalance" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.883341 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40afda2-a734-4619-a0e1-980d12f72780" containerName="swift-ring-rebalance" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.884012 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40afda2-a734-4619-a0e1-980d12f72780" containerName="swift-ring-rebalance" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.885895 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.888983 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.892585 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.916480 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-g5758"] Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.993895 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.994381 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.995663 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.996229 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l942\" (UniqueName: \"kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.996327 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.996564 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:03 crc kubenswrapper[4837]: I1001 10:22:03.996738 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097487 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097527 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097581 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097617 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097652 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l942\" (UniqueName: \"kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.097671 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.098301 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.098384 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.098433 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.105473 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.105896 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.106688 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.116332 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l942\" (UniqueName: \"kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942\") pod \"swift-ring-rebalance-debug-g5758\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:04 crc kubenswrapper[4837]: I1001 10:22:04.207284 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:05 crc kubenswrapper[4837]: I1001 10:22:05.245104 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-g5758"] Oct 01 10:22:05 crc kubenswrapper[4837]: W1001 10:22:05.262077 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda051a2d2_543a_4f52_a27f_07b9004a4926.slice/crio-a8603b74f1f2f069db31f2c9adcdc4fe2644ed2f28f2310d4009dd714ca2e0c3 WatchSource:0}: Error finding container a8603b74f1f2f069db31f2c9adcdc4fe2644ed2f28f2310d4009dd714ca2e0c3: Status 404 returned error can't find the container with id a8603b74f1f2f069db31f2c9adcdc4fe2644ed2f28f2310d4009dd714ca2e0c3 Oct 01 10:22:06 crc kubenswrapper[4837]: I1001 10:22:06.096116 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-g5758" event={"ID":"a051a2d2-543a-4f52-a27f-07b9004a4926","Type":"ContainerStarted","Data":"98b637cf56c5eecb58b7f4c00e8b899b24d6457f48c6640546a9adc914d5f52b"} Oct 01 10:22:06 crc kubenswrapper[4837]: I1001 10:22:06.096484 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-g5758" event={"ID":"a051a2d2-543a-4f52-a27f-07b9004a4926","Type":"ContainerStarted","Data":"a8603b74f1f2f069db31f2c9adcdc4fe2644ed2f28f2310d4009dd714ca2e0c3"} Oct 01 10:22:06 crc kubenswrapper[4837]: I1001 10:22:06.124819 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-g5758" podStartSLOduration=3.124795877 podStartE2EDuration="3.124795877s" podCreationTimestamp="2025-10-01 10:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:22:06.114339759 +0000 UTC m=+11782.955947224" watchObservedRunningTime="2025-10-01 10:22:06.124795877 +0000 UTC m=+11782.966403352" Oct 01 10:22:07 crc kubenswrapper[4837]: I1001 10:22:07.978998 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:07 crc kubenswrapper[4837]: I1001 10:22:07.987181 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:07 crc kubenswrapper[4837]: I1001 10:22:07.997787 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.081293 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2ffj\" (UniqueName: \"kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.081461 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.081522 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.183003 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.183084 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.183152 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2ffj\" (UniqueName: \"kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.183481 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.183605 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.209101 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2ffj\" (UniqueName: \"kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj\") pod \"redhat-marketplace-hlqpp\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:08 crc kubenswrapper[4837]: I1001 10:22:08.321611 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:09 crc kubenswrapper[4837]: I1001 10:22:09.009345 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:09 crc kubenswrapper[4837]: W1001 10:22:09.013496 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc84882fe_fc90_4763_8abc_4c469875c5a3.slice/crio-863b0d0a81981af200cc68e67e0507000e48649199b1f8a2eed638c0b93e6b70 WatchSource:0}: Error finding container 863b0d0a81981af200cc68e67e0507000e48649199b1f8a2eed638c0b93e6b70: Status 404 returned error can't find the container with id 863b0d0a81981af200cc68e67e0507000e48649199b1f8a2eed638c0b93e6b70 Oct 01 10:22:09 crc kubenswrapper[4837]: I1001 10:22:09.129276 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerStarted","Data":"863b0d0a81981af200cc68e67e0507000e48649199b1f8a2eed638c0b93e6b70"} Oct 01 10:22:10 crc kubenswrapper[4837]: I1001 10:22:10.143346 4837 generic.go:334] "Generic (PLEG): container finished" podID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerID="30a80886d6933c218b0a0b038f667dd236c25d62b33103298fdaf33b6adfa6e1" exitCode=0 Oct 01 10:22:10 crc kubenswrapper[4837]: I1001 10:22:10.143385 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerDied","Data":"30a80886d6933c218b0a0b038f667dd236c25d62b33103298fdaf33b6adfa6e1"} Oct 01 10:22:11 crc kubenswrapper[4837]: I1001 10:22:11.156832 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerStarted","Data":"aa05cd8fbae00caf79fe4646787b17f5fe3f8c3fc21d5be1a4f10bd781658479"} Oct 01 10:22:12 crc kubenswrapper[4837]: I1001 10:22:12.167075 4837 generic.go:334] "Generic (PLEG): container finished" podID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerID="aa05cd8fbae00caf79fe4646787b17f5fe3f8c3fc21d5be1a4f10bd781658479" exitCode=0 Oct 01 10:22:12 crc kubenswrapper[4837]: I1001 10:22:12.167184 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerDied","Data":"aa05cd8fbae00caf79fe4646787b17f5fe3f8c3fc21d5be1a4f10bd781658479"} Oct 01 10:22:13 crc kubenswrapper[4837]: I1001 10:22:13.180708 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerStarted","Data":"2cb6a94633ddd60c38b1df45d298e400d073d32f20cc07b369a8f1fefe7d3c60"} Oct 01 10:22:13 crc kubenswrapper[4837]: I1001 10:22:13.182523 4837 generic.go:334] "Generic (PLEG): container finished" podID="a051a2d2-543a-4f52-a27f-07b9004a4926" containerID="98b637cf56c5eecb58b7f4c00e8b899b24d6457f48c6640546a9adc914d5f52b" exitCode=0 Oct 01 10:22:13 crc kubenswrapper[4837]: I1001 10:22:13.182563 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-g5758" event={"ID":"a051a2d2-543a-4f52-a27f-07b9004a4926","Type":"ContainerDied","Data":"98b637cf56c5eecb58b7f4c00e8b899b24d6457f48c6640546a9adc914d5f52b"} Oct 01 10:22:13 crc kubenswrapper[4837]: I1001 10:22:13.213464 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hlqpp" podStartSLOduration=3.74218196 podStartE2EDuration="6.213438607s" podCreationTimestamp="2025-10-01 10:22:07 +0000 UTC" firstStartedPulling="2025-10-01 10:22:10.14721477 +0000 UTC m=+11786.988822225" lastFinishedPulling="2025-10-01 10:22:12.618471417 +0000 UTC m=+11789.460078872" observedRunningTime="2025-10-01 10:22:13.206581797 +0000 UTC m=+11790.048189252" watchObservedRunningTime="2025-10-01 10:22:13.213438607 +0000 UTC m=+11790.055046062" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.584182 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641317 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641445 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l942\" (UniqueName: \"kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641475 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641553 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641605 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641681 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.641767 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift\") pod \"a051a2d2-543a-4f52-a27f-07b9004a4926\" (UID: \"a051a2d2-543a-4f52-a27f-07b9004a4926\") " Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.643838 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.647144 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.649030 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-g5758"] Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.661914 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942" (OuterVolumeSpecName: "kube-api-access-8l942") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "kube-api-access-8l942". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.695390 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.696421 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts" (OuterVolumeSpecName: "scripts") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.696793 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-g5758"] Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.711468 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.734906 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a051a2d2-543a-4f52-a27f-07b9004a4926" (UID: "a051a2d2-543a-4f52-a27f-07b9004a4926"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.747996 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l942\" (UniqueName: \"kubernetes.io/projected/a051a2d2-543a-4f52-a27f-07b9004a4926-kube-api-access-8l942\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748029 4837 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748039 4837 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748049 4837 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748056 4837 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a051a2d2-543a-4f52-a27f-07b9004a4926-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748064 4837 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a051a2d2-543a-4f52-a27f-07b9004a4926-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.748072 4837 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a051a2d2-543a-4f52-a27f-07b9004a4926-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:15 crc kubenswrapper[4837]: I1001 10:22:15.826360 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a051a2d2-543a-4f52-a27f-07b9004a4926" path="/var/lib/kubelet/pods/a051a2d2-543a-4f52-a27f-07b9004a4926/volumes" Oct 01 10:22:16 crc kubenswrapper[4837]: I1001 10:22:16.224489 4837 scope.go:117] "RemoveContainer" containerID="98b637cf56c5eecb58b7f4c00e8b899b24d6457f48c6640546a9adc914d5f52b" Oct 01 10:22:16 crc kubenswrapper[4837]: I1001 10:22:16.224527 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-g5758" Oct 01 10:22:18 crc kubenswrapper[4837]: I1001 10:22:18.322265 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:18 crc kubenswrapper[4837]: I1001 10:22:18.322528 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:18 crc kubenswrapper[4837]: I1001 10:22:18.373318 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:18 crc kubenswrapper[4837]: I1001 10:22:18.514855 4837 scope.go:117] "RemoveContainer" containerID="96ac6a2aee16a0b3a3e31b79f354a25e57522dc37ec5c63d1b2d182d6f9a1263" Oct 01 10:22:18 crc kubenswrapper[4837]: I1001 10:22:18.544677 4837 scope.go:117] "RemoveContainer" containerID="3906f614001dd109df93dba78087c0c78f6c0aa41c3dadabcd220e8ea90a1504" Oct 01 10:22:19 crc kubenswrapper[4837]: I1001 10:22:19.315453 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:19 crc kubenswrapper[4837]: I1001 10:22:19.366573 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:21 crc kubenswrapper[4837]: I1001 10:22:21.287168 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hlqpp" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="registry-server" containerID="cri-o://2cb6a94633ddd60c38b1df45d298e400d073d32f20cc07b369a8f1fefe7d3c60" gracePeriod=2 Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.299583 4837 generic.go:334] "Generic (PLEG): container finished" podID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerID="2cb6a94633ddd60c38b1df45d298e400d073d32f20cc07b369a8f1fefe7d3c60" exitCode=0 Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.299688 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerDied","Data":"2cb6a94633ddd60c38b1df45d298e400d073d32f20cc07b369a8f1fefe7d3c60"} Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.818342 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.909971 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2ffj\" (UniqueName: \"kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj\") pod \"c84882fe-fc90-4763-8abc-4c469875c5a3\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.910272 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content\") pod \"c84882fe-fc90-4763-8abc-4c469875c5a3\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.910315 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities\") pod \"c84882fe-fc90-4763-8abc-4c469875c5a3\" (UID: \"c84882fe-fc90-4763-8abc-4c469875c5a3\") " Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.913852 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities" (OuterVolumeSpecName: "utilities") pod "c84882fe-fc90-4763-8abc-4c469875c5a3" (UID: "c84882fe-fc90-4763-8abc-4c469875c5a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.918073 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj" (OuterVolumeSpecName: "kube-api-access-s2ffj") pod "c84882fe-fc90-4763-8abc-4c469875c5a3" (UID: "c84882fe-fc90-4763-8abc-4c469875c5a3"). InnerVolumeSpecName "kube-api-access-s2ffj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:22:22 crc kubenswrapper[4837]: I1001 10:22:22.948082 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c84882fe-fc90-4763-8abc-4c469875c5a3" (UID: "c84882fe-fc90-4763-8abc-4c469875c5a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.013002 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.013035 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c84882fe-fc90-4763-8abc-4c469875c5a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.013044 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2ffj\" (UniqueName: \"kubernetes.io/projected/c84882fe-fc90-4763-8abc-4c469875c5a3-kube-api-access-s2ffj\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.079010 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.079058 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.321673 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlqpp" event={"ID":"c84882fe-fc90-4763-8abc-4c469875c5a3","Type":"ContainerDied","Data":"863b0d0a81981af200cc68e67e0507000e48649199b1f8a2eed638c0b93e6b70"} Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.321741 4837 scope.go:117] "RemoveContainer" containerID="2cb6a94633ddd60c38b1df45d298e400d073d32f20cc07b369a8f1fefe7d3c60" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.321767 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlqpp" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.360444 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.361870 4837 scope.go:117] "RemoveContainer" containerID="aa05cd8fbae00caf79fe4646787b17f5fe3f8c3fc21d5be1a4f10bd781658479" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.373857 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlqpp"] Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.394285 4837 scope.go:117] "RemoveContainer" containerID="30a80886d6933c218b0a0b038f667dd236c25d62b33103298fdaf33b6adfa6e1" Oct 01 10:22:23 crc kubenswrapper[4837]: I1001 10:22:23.828285 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" path="/var/lib/kubelet/pods/c84882fe-fc90-4763-8abc-4c469875c5a3/volumes" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.639809 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:22:30 crc kubenswrapper[4837]: E1001 10:22:30.640948 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="registry-server" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.640965 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="registry-server" Oct 01 10:22:30 crc kubenswrapper[4837]: E1001 10:22:30.640993 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a051a2d2-543a-4f52-a27f-07b9004a4926" containerName="swift-ring-rebalance" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.641001 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="a051a2d2-543a-4f52-a27f-07b9004a4926" containerName="swift-ring-rebalance" Oct 01 10:22:30 crc kubenswrapper[4837]: E1001 10:22:30.641027 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="extract-utilities" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.641035 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="extract-utilities" Oct 01 10:22:30 crc kubenswrapper[4837]: E1001 10:22:30.641064 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="extract-content" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.641072 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="extract-content" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.641331 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="a051a2d2-543a-4f52-a27f-07b9004a4926" containerName="swift-ring-rebalance" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.641374 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="c84882fe-fc90-4763-8abc-4c469875c5a3" containerName="registry-server" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.642251 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.645812 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.646006 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.646538 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cz4sn" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.648899 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.653906 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769735 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769824 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769846 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769882 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769901 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769920 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769942 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz92l\" (UniqueName: \"kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769959 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.769975 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.872193 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.872643 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.872786 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.872897 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.872998 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.873110 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz92l\" (UniqueName: \"kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.873202 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.873305 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.873379 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.873562 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.874114 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.876204 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.877072 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.878065 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.878612 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.880163 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.884001 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.898496 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz92l\" (UniqueName: \"kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.914484 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " pod="openstack/tempest-tests-tempest" Oct 01 10:22:30 crc kubenswrapper[4837]: I1001 10:22:30.973033 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:22:31 crc kubenswrapper[4837]: I1001 10:22:31.670245 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:22:32 crc kubenswrapper[4837]: I1001 10:22:32.416090 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"62fa4c75-1944-4f6f-8efe-8463b065c5c5","Type":"ContainerStarted","Data":"e0e51fd56d35bf27d3c4af2d6e5f58ce916d11bf307097afa6316dd74b85ff7a"} Oct 01 10:22:53 crc kubenswrapper[4837]: I1001 10:22:53.079166 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:22:53 crc kubenswrapper[4837]: I1001 10:22:53.079814 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.524902 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.528297 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.538713 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.544975 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wf4f\" (UniqueName: \"kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.545033 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.545149 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.646815 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wf4f\" (UniqueName: \"kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.646859 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.646943 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.647455 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.647837 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.667364 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wf4f\" (UniqueName: \"kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f\") pod \"redhat-operators-tqm9x\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:16 crc kubenswrapper[4837]: I1001 10:23:16.868301 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:18 crc kubenswrapper[4837]: I1001 10:23:18.700877 4837 scope.go:117] "RemoveContainer" containerID="ec5028f8f737edf2d9958b3c2c288c4bddbc88d69ca6de0230ce56a186fa85bf" Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.079770 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.080315 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.080369 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.081354 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.081422 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" gracePeriod=600 Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.991576 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" exitCode=0 Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.991662 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce"} Oct 01 10:23:23 crc kubenswrapper[4837]: I1001 10:23:23.991889 4837 scope.go:117] "RemoveContainer" containerID="175d232c0edca5e070b31d15c701afc3380dea84fa349d315b010654361aa995" Oct 01 10:23:25 crc kubenswrapper[4837]: E1001 10:23:25.439286 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:23:25 crc kubenswrapper[4837]: E1001 10:23:25.514539 4837 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:a0eac564d779a7eaac46c9816bff261a" Oct 01 10:23:25 crc kubenswrapper[4837]: E1001 10:23:25.514928 4837 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:a0eac564d779a7eaac46c9816bff261a" Oct 01 10:23:25 crc kubenswrapper[4837]: E1001 10:23:25.515354 4837 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:a0eac564d779a7eaac46c9816bff261a,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gz92l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(62fa4c75-1944-4f6f-8efe-8463b065c5c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 10:23:25 crc kubenswrapper[4837]: E1001 10:23:25.516683 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" Oct 01 10:23:26 crc kubenswrapper[4837]: I1001 10:23:26.044784 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:23:26 crc kubenswrapper[4837]: E1001 10:23:26.045489 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:23:26 crc kubenswrapper[4837]: E1001 10:23:26.046219 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:a0eac564d779a7eaac46c9816bff261a\\\"\"" pod="openstack/tempest-tests-tempest" podUID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" Oct 01 10:23:26 crc kubenswrapper[4837]: I1001 10:23:26.277847 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:27 crc kubenswrapper[4837]: I1001 10:23:27.063070 4837 generic.go:334] "Generic (PLEG): container finished" podID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerID="50e1259a15f14b243c1f305099dee767391939f415554f6e0ae8ab06160adb47" exitCode=0 Oct 01 10:23:27 crc kubenswrapper[4837]: I1001 10:23:27.063280 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerDied","Data":"50e1259a15f14b243c1f305099dee767391939f415554f6e0ae8ab06160adb47"} Oct 01 10:23:27 crc kubenswrapper[4837]: I1001 10:23:27.063303 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerStarted","Data":"75e1c61c46cf136288ff48f0ee48d165ab66a851214415eea0b498e9aeb98f3b"} Oct 01 10:23:27 crc kubenswrapper[4837]: I1001 10:23:27.065391 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:23:28 crc kubenswrapper[4837]: I1001 10:23:28.083567 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerStarted","Data":"4a2f5b8b96c1321befed54612ee10898911dcd8dcecdda9049a83761d3667209"} Oct 01 10:23:34 crc kubenswrapper[4837]: I1001 10:23:34.153915 4837 generic.go:334] "Generic (PLEG): container finished" podID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerID="4a2f5b8b96c1321befed54612ee10898911dcd8dcecdda9049a83761d3667209" exitCode=0 Oct 01 10:23:34 crc kubenswrapper[4837]: I1001 10:23:34.153976 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerDied","Data":"4a2f5b8b96c1321befed54612ee10898911dcd8dcecdda9049a83761d3667209"} Oct 01 10:23:35 crc kubenswrapper[4837]: I1001 10:23:35.167754 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerStarted","Data":"b24b922280211a8fa3aab8316303c3947a7b9005ec2e63a47a4b91f0b914ccee"} Oct 01 10:23:35 crc kubenswrapper[4837]: I1001 10:23:35.197507 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tqm9x" podStartSLOduration=11.456577109 podStartE2EDuration="19.197485535s" podCreationTimestamp="2025-10-01 10:23:16 +0000 UTC" firstStartedPulling="2025-10-01 10:23:27.065151616 +0000 UTC m=+11863.906759071" lastFinishedPulling="2025-10-01 10:23:34.806060042 +0000 UTC m=+11871.647667497" observedRunningTime="2025-10-01 10:23:35.186449952 +0000 UTC m=+11872.028057427" watchObservedRunningTime="2025-10-01 10:23:35.197485535 +0000 UTC m=+11872.039093010" Oct 01 10:23:36 crc kubenswrapper[4837]: I1001 10:23:36.873818 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:36 crc kubenswrapper[4837]: I1001 10:23:36.874143 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:37 crc kubenswrapper[4837]: I1001 10:23:37.940533 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tqm9x" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="registry-server" probeResult="failure" output=< Oct 01 10:23:37 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:23:37 crc kubenswrapper[4837]: > Oct 01 10:23:39 crc kubenswrapper[4837]: I1001 10:23:39.096953 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 10:23:40 crc kubenswrapper[4837]: I1001 10:23:40.816287 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:23:40 crc kubenswrapper[4837]: E1001 10:23:40.816820 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:23:41 crc kubenswrapper[4837]: I1001 10:23:41.248813 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"62fa4c75-1944-4f6f-8efe-8463b065c5c5","Type":"ContainerStarted","Data":"047cf8eb1f04848fdf1eb08f432a6af8d7c90f7ec3ff78c5774c2381d196b016"} Oct 01 10:23:41 crc kubenswrapper[4837]: I1001 10:23:41.277000 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.863562239 podStartE2EDuration="1m12.27697973s" podCreationTimestamp="2025-10-01 10:22:29 +0000 UTC" firstStartedPulling="2025-10-01 10:22:31.680790718 +0000 UTC m=+11808.522398173" lastFinishedPulling="2025-10-01 10:23:39.094208209 +0000 UTC m=+11875.935815664" observedRunningTime="2025-10-01 10:23:41.269436873 +0000 UTC m=+11878.111044328" watchObservedRunningTime="2025-10-01 10:23:41.27697973 +0000 UTC m=+11878.118587185" Oct 01 10:23:46 crc kubenswrapper[4837]: I1001 10:23:46.927376 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:47 crc kubenswrapper[4837]: I1001 10:23:47.000181 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:47 crc kubenswrapper[4837]: I1001 10:23:47.729966 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:48 crc kubenswrapper[4837]: I1001 10:23:48.336494 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tqm9x" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="registry-server" containerID="cri-o://b24b922280211a8fa3aab8316303c3947a7b9005ec2e63a47a4b91f0b914ccee" gracePeriod=2 Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.350124 4837 generic.go:334] "Generic (PLEG): container finished" podID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerID="b24b922280211a8fa3aab8316303c3947a7b9005ec2e63a47a4b91f0b914ccee" exitCode=0 Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.350185 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerDied","Data":"b24b922280211a8fa3aab8316303c3947a7b9005ec2e63a47a4b91f0b914ccee"} Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.749176 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.850550 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities\") pod \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.851173 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content\") pod \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.851340 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wf4f\" (UniqueName: \"kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f\") pod \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\" (UID: \"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d\") " Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.851467 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities" (OuterVolumeSpecName: "utilities") pod "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" (UID: "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.852096 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.858616 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f" (OuterVolumeSpecName: "kube-api-access-6wf4f") pod "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" (UID: "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d"). InnerVolumeSpecName "kube-api-access-6wf4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.954214 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wf4f\" (UniqueName: \"kubernetes.io/projected/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-kube-api-access-6wf4f\") on node \"crc\" DevicePath \"\"" Oct 01 10:23:49 crc kubenswrapper[4837]: I1001 10:23:49.954370 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" (UID: "b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.056090 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.362597 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqm9x" event={"ID":"b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d","Type":"ContainerDied","Data":"75e1c61c46cf136288ff48f0ee48d165ab66a851214415eea0b498e9aeb98f3b"} Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.362651 4837 scope.go:117] "RemoveContainer" containerID="b24b922280211a8fa3aab8316303c3947a7b9005ec2e63a47a4b91f0b914ccee" Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.362663 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqm9x" Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.385650 4837 scope.go:117] "RemoveContainer" containerID="4a2f5b8b96c1321befed54612ee10898911dcd8dcecdda9049a83761d3667209" Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.406739 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.417186 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tqm9x"] Oct 01 10:23:50 crc kubenswrapper[4837]: I1001 10:23:50.430615 4837 scope.go:117] "RemoveContainer" containerID="50e1259a15f14b243c1f305099dee767391939f415554f6e0ae8ab06160adb47" Oct 01 10:23:51 crc kubenswrapper[4837]: I1001 10:23:51.832150 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" path="/var/lib/kubelet/pods/b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d/volumes" Oct 01 10:23:52 crc kubenswrapper[4837]: I1001 10:23:52.816521 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:23:52 crc kubenswrapper[4837]: E1001 10:23:52.816813 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:24:04 crc kubenswrapper[4837]: I1001 10:24:04.816301 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:24:04 crc kubenswrapper[4837]: E1001 10:24:04.817152 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:24:16 crc kubenswrapper[4837]: I1001 10:24:16.815856 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:24:16 crc kubenswrapper[4837]: E1001 10:24:16.816729 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:24:25 crc kubenswrapper[4837]: I1001 10:24:25.583609 4837 scope.go:117] "RemoveContainer" containerID="201acb359643a6e002344bc1c84223c2e322389d2928e530c1630b61e30828aa" Oct 01 10:24:31 crc kubenswrapper[4837]: I1001 10:24:31.817015 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:24:31 crc kubenswrapper[4837]: E1001 10:24:31.817881 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:24:44 crc kubenswrapper[4837]: I1001 10:24:44.815679 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:24:44 crc kubenswrapper[4837]: E1001 10:24:44.816404 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:24:56 crc kubenswrapper[4837]: I1001 10:24:56.816883 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:24:56 crc kubenswrapper[4837]: E1001 10:24:56.817751 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:25:10 crc kubenswrapper[4837]: I1001 10:25:10.816506 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:25:10 crc kubenswrapper[4837]: E1001 10:25:10.817349 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:25:23 crc kubenswrapper[4837]: I1001 10:25:23.822935 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:25:23 crc kubenswrapper[4837]: E1001 10:25:23.823523 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:25:25 crc kubenswrapper[4837]: I1001 10:25:25.696017 4837 scope.go:117] "RemoveContainer" containerID="f2d8273b78ac8fcbf6f44debcdc110989cf1c42bad270d86beee2b94a0986bc4" Oct 01 10:25:25 crc kubenswrapper[4837]: I1001 10:25:25.743176 4837 scope.go:117] "RemoveContainer" containerID="9d0cc08c4d77c4d6a0db5327ed0c4973d0ac27a6104936256e5a658300645b53" Oct 01 10:25:35 crc kubenswrapper[4837]: I1001 10:25:35.815880 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:25:35 crc kubenswrapper[4837]: E1001 10:25:35.816818 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:25:50 crc kubenswrapper[4837]: I1001 10:25:50.817190 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:25:50 crc kubenswrapper[4837]: E1001 10:25:50.818107 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:26:01 crc kubenswrapper[4837]: I1001 10:26:01.816919 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:26:01 crc kubenswrapper[4837]: E1001 10:26:01.817826 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:26:15 crc kubenswrapper[4837]: I1001 10:26:15.816734 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:26:15 crc kubenswrapper[4837]: E1001 10:26:15.817317 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:26:27 crc kubenswrapper[4837]: I1001 10:26:27.816442 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:26:27 crc kubenswrapper[4837]: E1001 10:26:27.817383 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:26:39 crc kubenswrapper[4837]: I1001 10:26:39.816377 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:26:39 crc kubenswrapper[4837]: E1001 10:26:39.817298 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:26:53 crc kubenswrapper[4837]: I1001 10:26:53.878349 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:26:53 crc kubenswrapper[4837]: E1001 10:26:53.879290 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:27:05 crc kubenswrapper[4837]: I1001 10:27:05.816596 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:27:05 crc kubenswrapper[4837]: E1001 10:27:05.817476 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:27:18 crc kubenswrapper[4837]: I1001 10:27:18.816341 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:27:18 crc kubenswrapper[4837]: E1001 10:27:18.817299 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:27:33 crc kubenswrapper[4837]: I1001 10:27:33.823625 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:27:33 crc kubenswrapper[4837]: E1001 10:27:33.824401 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:27:45 crc kubenswrapper[4837]: I1001 10:27:45.816517 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:27:45 crc kubenswrapper[4837]: E1001 10:27:45.817139 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:27:58 crc kubenswrapper[4837]: I1001 10:27:58.815992 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:27:58 crc kubenswrapper[4837]: E1001 10:27:58.816671 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.283621 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:12 crc kubenswrapper[4837]: E1001 10:28:12.284943 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="registry-server" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.284967 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="registry-server" Oct 01 10:28:12 crc kubenswrapper[4837]: E1001 10:28:12.284980 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="extract-content" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.284986 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="extract-content" Oct 01 10:28:12 crc kubenswrapper[4837]: E1001 10:28:12.285000 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="extract-utilities" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.285007 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="extract-utilities" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.285234 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b077aabb-1fb0-4fd6-92a7-9b5a6ba4cb1d" containerName="registry-server" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.287172 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.306081 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.416968 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.417046 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tftgz\" (UniqueName: \"kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.417205 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.519017 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.519083 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tftgz\" (UniqueName: \"kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.519114 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.519584 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.519629 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.548026 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tftgz\" (UniqueName: \"kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz\") pod \"community-operators-tnhj2\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.605206 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:12 crc kubenswrapper[4837]: I1001 10:28:12.816451 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:28:12 crc kubenswrapper[4837]: E1001 10:28:12.816941 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:28:13 crc kubenswrapper[4837]: I1001 10:28:13.744324 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.084112 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.087181 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.125310 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.165674 4837 generic.go:334] "Generic (PLEG): container finished" podID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerID="73c9df36a851040a6db20ec4b2d4d2e126146e6dec94e3c1637a0a3eab74805f" exitCode=0 Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.165739 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerDied","Data":"73c9df36a851040a6db20ec4b2d4d2e126146e6dec94e3c1637a0a3eab74805f"} Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.165769 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerStarted","Data":"ce4d638559f5cf0f60daaeffe615961747177c337f793125e3e257fff38c087f"} Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.270872 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.270997 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpj99\" (UniqueName: \"kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.271145 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.372845 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.373042 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.373125 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpj99\" (UniqueName: \"kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.373969 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.374548 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.397053 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpj99\" (UniqueName: \"kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99\") pod \"certified-operators-msvz6\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:14 crc kubenswrapper[4837]: I1001 10:28:14.419016 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:15 crc kubenswrapper[4837]: I1001 10:28:15.402640 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:16 crc kubenswrapper[4837]: I1001 10:28:16.188490 4837 generic.go:334] "Generic (PLEG): container finished" podID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerID="e67b1bcd39e91d8c1bb870690da0fda451c080b772b38e43e9aa2f12c8b72665" exitCode=0 Oct 01 10:28:16 crc kubenswrapper[4837]: I1001 10:28:16.188543 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerDied","Data":"e67b1bcd39e91d8c1bb870690da0fda451c080b772b38e43e9aa2f12c8b72665"} Oct 01 10:28:16 crc kubenswrapper[4837]: I1001 10:28:16.189277 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerStarted","Data":"bc5a01f2f5657c0e304be7f8075fec68a0f4045a4649593237a25385d067bbd3"} Oct 01 10:28:16 crc kubenswrapper[4837]: I1001 10:28:16.192222 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerStarted","Data":"69665c3d0b54e9db058ef175b7139c42c2e0f8d39c8f6f356ebde59dd71b7f77"} Oct 01 10:28:19 crc kubenswrapper[4837]: I1001 10:28:19.225787 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerStarted","Data":"e973361b089ad75b22300dc662e18b345749f90de526e60216cb96e31594bef9"} Oct 01 10:28:21 crc kubenswrapper[4837]: I1001 10:28:21.246503 4837 generic.go:334] "Generic (PLEG): container finished" podID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerID="69665c3d0b54e9db058ef175b7139c42c2e0f8d39c8f6f356ebde59dd71b7f77" exitCode=0 Oct 01 10:28:21 crc kubenswrapper[4837]: I1001 10:28:21.246543 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerDied","Data":"69665c3d0b54e9db058ef175b7139c42c2e0f8d39c8f6f356ebde59dd71b7f77"} Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.257975 4837 generic.go:334] "Generic (PLEG): container finished" podID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerID="e973361b089ad75b22300dc662e18b345749f90de526e60216cb96e31594bef9" exitCode=0 Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.258071 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerDied","Data":"e973361b089ad75b22300dc662e18b345749f90de526e60216cb96e31594bef9"} Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.263251 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerStarted","Data":"e4e2b70523b9c15d8afc008d6081abc43b15a0fceeed634d49b10c096dd515f5"} Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.304396 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tnhj2" podStartSLOduration=2.812988595 podStartE2EDuration="10.304378937s" podCreationTimestamp="2025-10-01 10:28:12 +0000 UTC" firstStartedPulling="2025-10-01 10:28:14.167622519 +0000 UTC m=+12151.009229974" lastFinishedPulling="2025-10-01 10:28:21.659012861 +0000 UTC m=+12158.500620316" observedRunningTime="2025-10-01 10:28:22.300978474 +0000 UTC m=+12159.142585939" watchObservedRunningTime="2025-10-01 10:28:22.304378937 +0000 UTC m=+12159.145986392" Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.605962 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:22 crc kubenswrapper[4837]: I1001 10:28:22.606013 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:23 crc kubenswrapper[4837]: I1001 10:28:23.277401 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerStarted","Data":"24975b24eab243c5115553b3a9641030dc9cec376b49e1e7e90c39721ff6041a"} Oct 01 10:28:23 crc kubenswrapper[4837]: I1001 10:28:23.303662 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-msvz6" podStartSLOduration=2.768606507 podStartE2EDuration="9.303636239s" podCreationTimestamp="2025-10-01 10:28:14 +0000 UTC" firstStartedPulling="2025-10-01 10:28:16.191129239 +0000 UTC m=+12153.032736694" lastFinishedPulling="2025-10-01 10:28:22.726158981 +0000 UTC m=+12159.567766426" observedRunningTime="2025-10-01 10:28:23.296571674 +0000 UTC m=+12160.138179129" watchObservedRunningTime="2025-10-01 10:28:23.303636239 +0000 UTC m=+12160.145243694" Oct 01 10:28:23 crc kubenswrapper[4837]: I1001 10:28:23.662392 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tnhj2" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" probeResult="failure" output=< Oct 01 10:28:23 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:28:23 crc kubenswrapper[4837]: > Oct 01 10:28:24 crc kubenswrapper[4837]: I1001 10:28:24.419764 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:24 crc kubenswrapper[4837]: I1001 10:28:24.420035 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:24 crc kubenswrapper[4837]: I1001 10:28:24.816459 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:28:25 crc kubenswrapper[4837]: I1001 10:28:25.300989 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504"} Oct 01 10:28:25 crc kubenswrapper[4837]: I1001 10:28:25.482850 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-msvz6" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" probeResult="failure" output=< Oct 01 10:28:25 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:28:25 crc kubenswrapper[4837]: > Oct 01 10:28:33 crc kubenswrapper[4837]: I1001 10:28:33.658987 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tnhj2" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" probeResult="failure" output=< Oct 01 10:28:33 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:28:33 crc kubenswrapper[4837]: > Oct 01 10:28:35 crc kubenswrapper[4837]: I1001 10:28:35.469474 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-msvz6" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" probeResult="failure" output=< Oct 01 10:28:35 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:28:35 crc kubenswrapper[4837]: > Oct 01 10:28:42 crc kubenswrapper[4837]: I1001 10:28:42.669768 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:42 crc kubenswrapper[4837]: I1001 10:28:42.731962 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:43 crc kubenswrapper[4837]: I1001 10:28:43.488443 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:44 crc kubenswrapper[4837]: I1001 10:28:44.490099 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:44 crc kubenswrapper[4837]: I1001 10:28:44.523783 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tnhj2" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" containerID="cri-o://e4e2b70523b9c15d8afc008d6081abc43b15a0fceeed634d49b10c096dd515f5" gracePeriod=2 Oct 01 10:28:44 crc kubenswrapper[4837]: I1001 10:28:44.540379 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:45 crc kubenswrapper[4837]: I1001 10:28:45.536065 4837 generic.go:334] "Generic (PLEG): container finished" podID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerID="e4e2b70523b9c15d8afc008d6081abc43b15a0fceeed634d49b10c096dd515f5" exitCode=0 Oct 01 10:28:45 crc kubenswrapper[4837]: I1001 10:28:45.536359 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerDied","Data":"e4e2b70523b9c15d8afc008d6081abc43b15a0fceeed634d49b10c096dd515f5"} Oct 01 10:28:45 crc kubenswrapper[4837]: I1001 10:28:45.885862 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:45 crc kubenswrapper[4837]: I1001 10:28:45.886306 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-msvz6" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" containerID="cri-o://24975b24eab243c5115553b3a9641030dc9cec376b49e1e7e90c39721ff6041a" gracePeriod=2 Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.595005 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.595367 4837 generic.go:334] "Generic (PLEG): container finished" podID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerID="24975b24eab243c5115553b3a9641030dc9cec376b49e1e7e90c39721ff6041a" exitCode=0 Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.595473 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerDied","Data":"24975b24eab243c5115553b3a9641030dc9cec376b49e1e7e90c39721ff6041a"} Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.614764 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnhj2" event={"ID":"4a6728eb-259d-4dd1-b224-f1d7546f9f87","Type":"ContainerDied","Data":"ce4d638559f5cf0f60daaeffe615961747177c337f793125e3e257fff38c087f"} Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.614809 4837 scope.go:117] "RemoveContainer" containerID="e4e2b70523b9c15d8afc008d6081abc43b15a0fceeed634d49b10c096dd515f5" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.614948 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnhj2" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.667095 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content\") pod \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.667317 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities\") pod \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.667416 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tftgz\" (UniqueName: \"kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz\") pod \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\" (UID: \"4a6728eb-259d-4dd1-b224-f1d7546f9f87\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.669180 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities" (OuterVolumeSpecName: "utilities") pod "4a6728eb-259d-4dd1-b224-f1d7546f9f87" (UID: "4a6728eb-259d-4dd1-b224-f1d7546f9f87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.692048 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz" (OuterVolumeSpecName: "kube-api-access-tftgz") pod "4a6728eb-259d-4dd1-b224-f1d7546f9f87" (UID: "4a6728eb-259d-4dd1-b224-f1d7546f9f87"). InnerVolumeSpecName "kube-api-access-tftgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.715843 4837 scope.go:117] "RemoveContainer" containerID="69665c3d0b54e9db058ef175b7139c42c2e0f8d39c8f6f356ebde59dd71b7f77" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.748749 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a6728eb-259d-4dd1-b224-f1d7546f9f87" (UID: "4a6728eb-259d-4dd1-b224-f1d7546f9f87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.787362 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.787383 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tftgz\" (UniqueName: \"kubernetes.io/projected/4a6728eb-259d-4dd1-b224-f1d7546f9f87-kube-api-access-tftgz\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.787396 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a6728eb-259d-4dd1-b224-f1d7546f9f87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.818679 4837 scope.go:117] "RemoveContainer" containerID="73c9df36a851040a6db20ec4b2d4d2e126146e6dec94e3c1637a0a3eab74805f" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.951423 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:46.969825 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tnhj2"] Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.467768 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.609910 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content\") pod \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.610067 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpj99\" (UniqueName: \"kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99\") pod \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.610098 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities\") pod \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\" (UID: \"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4\") " Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.611385 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities" (OuterVolumeSpecName: "utilities") pod "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" (UID: "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.615904 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99" (OuterVolumeSpecName: "kube-api-access-zpj99") pod "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" (UID: "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4"). InnerVolumeSpecName "kube-api-access-zpj99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.660081 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-msvz6" event={"ID":"3424c24c-8db6-4bae-a84d-3c3ed3da8cf4","Type":"ContainerDied","Data":"bc5a01f2f5657c0e304be7f8075fec68a0f4045a4649593237a25385d067bbd3"} Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.660138 4837 scope.go:117] "RemoveContainer" containerID="24975b24eab243c5115553b3a9641030dc9cec376b49e1e7e90c39721ff6041a" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.660284 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-msvz6" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.712273 4837 scope.go:117] "RemoveContainer" containerID="e973361b089ad75b22300dc662e18b345749f90de526e60216cb96e31594bef9" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.712333 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpj99\" (UniqueName: \"kubernetes.io/projected/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-kube-api-access-zpj99\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.712356 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.713066 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" (UID: "3424c24c-8db6-4bae-a84d-3c3ed3da8cf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.743623 4837 scope.go:117] "RemoveContainer" containerID="e67b1bcd39e91d8c1bb870690da0fda451c080b772b38e43e9aa2f12c8b72665" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.813747 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.829592 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" path="/var/lib/kubelet/pods/4a6728eb-259d-4dd1-b224-f1d7546f9f87/volumes" Oct 01 10:28:47 crc kubenswrapper[4837]: I1001 10:28:47.994060 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:48 crc kubenswrapper[4837]: I1001 10:28:48.005531 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-msvz6"] Oct 01 10:28:49 crc kubenswrapper[4837]: I1001 10:28:49.827769 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" path="/var/lib/kubelet/pods/3424c24c-8db6-4bae-a84d-3c3ed3da8cf4/volumes" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.173631 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f"] Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174817 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174836 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174847 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174855 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174873 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174881 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174908 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174915 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174943 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174950 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4837]: E1001 10:30:00.174976 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.174985 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.175232 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="3424c24c-8db6-4bae-a84d-3c3ed3da8cf4" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.175263 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a6728eb-259d-4dd1-b224-f1d7546f9f87" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.176157 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.178801 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.179128 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.186423 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f"] Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.215889 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2cz\" (UniqueName: \"kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.216263 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.216459 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.320550 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.320685 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2cz\" (UniqueName: \"kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.320755 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.321744 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.332898 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.336927 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2cz\" (UniqueName: \"kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz\") pod \"collect-profiles-29321910-4w77f\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:00 crc kubenswrapper[4837]: I1001 10:30:00.506003 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:01 crc kubenswrapper[4837]: I1001 10:30:01.498378 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f"] Oct 01 10:30:02 crc kubenswrapper[4837]: I1001 10:30:02.446431 4837 generic.go:334] "Generic (PLEG): container finished" podID="cec3907b-97c0-47a7-b1c2-27ee5fd92148" containerID="f91af1acfa6c402328b46f659ad9c7de303d4b01042d5644ce1135c05a344f52" exitCode=0 Oct 01 10:30:02 crc kubenswrapper[4837]: I1001 10:30:02.446534 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" event={"ID":"cec3907b-97c0-47a7-b1c2-27ee5fd92148","Type":"ContainerDied","Data":"f91af1acfa6c402328b46f659ad9c7de303d4b01042d5644ce1135c05a344f52"} Oct 01 10:30:02 crc kubenswrapper[4837]: I1001 10:30:02.446734 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" event={"ID":"cec3907b-97c0-47a7-b1c2-27ee5fd92148","Type":"ContainerStarted","Data":"db50b695e5e05fb7cfed4ba91aba6a87ee34d3d199e5dae03db464aa2df59615"} Oct 01 10:30:04 crc kubenswrapper[4837]: I1001 10:30:04.880545 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.034838 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume\") pod \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.035183 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume\") pod \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.035283 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj2cz\" (UniqueName: \"kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz\") pod \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\" (UID: \"cec3907b-97c0-47a7-b1c2-27ee5fd92148\") " Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.035675 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume" (OuterVolumeSpecName: "config-volume") pod "cec3907b-97c0-47a7-b1c2-27ee5fd92148" (UID: "cec3907b-97c0-47a7-b1c2-27ee5fd92148"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.035815 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cec3907b-97c0-47a7-b1c2-27ee5fd92148-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.050374 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cec3907b-97c0-47a7-b1c2-27ee5fd92148" (UID: "cec3907b-97c0-47a7-b1c2-27ee5fd92148"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.053786 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz" (OuterVolumeSpecName: "kube-api-access-nj2cz") pod "cec3907b-97c0-47a7-b1c2-27ee5fd92148" (UID: "cec3907b-97c0-47a7-b1c2-27ee5fd92148"). InnerVolumeSpecName "kube-api-access-nj2cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.138464 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj2cz\" (UniqueName: \"kubernetes.io/projected/cec3907b-97c0-47a7-b1c2-27ee5fd92148-kube-api-access-nj2cz\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.138504 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cec3907b-97c0-47a7-b1c2-27ee5fd92148-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.481724 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" event={"ID":"cec3907b-97c0-47a7-b1c2-27ee5fd92148","Type":"ContainerDied","Data":"db50b695e5e05fb7cfed4ba91aba6a87ee34d3d199e5dae03db464aa2df59615"} Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.481769 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db50b695e5e05fb7cfed4ba91aba6a87ee34d3d199e5dae03db464aa2df59615" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.481831 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-4w77f" Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.957111 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h"] Oct 01 10:30:05 crc kubenswrapper[4837]: I1001 10:30:05.972868 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-vzb2h"] Oct 01 10:30:07 crc kubenswrapper[4837]: I1001 10:30:07.833175 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0ead52-9298-441f-bca2-7fadf774448b" path="/var/lib/kubelet/pods/5b0ead52-9298-441f-bca2-7fadf774448b/volumes" Oct 01 10:30:26 crc kubenswrapper[4837]: I1001 10:30:26.004075 4837 scope.go:117] "RemoveContainer" containerID="2847180ad0b9adc1a6d26ad7a34cff906738ccbb64a3cd6686c7889fb5e172c0" Oct 01 10:30:53 crc kubenswrapper[4837]: I1001 10:30:53.079281 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:30:53 crc kubenswrapper[4837]: I1001 10:30:53.079844 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:31:23 crc kubenswrapper[4837]: I1001 10:31:23.079770 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:31:23 crc kubenswrapper[4837]: I1001 10:31:23.080376 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.079106 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.080059 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.080109 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.081331 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.081405 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504" gracePeriod=600 Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.787198 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504" exitCode=0 Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.788821 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504"} Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.788891 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821"} Oct 01 10:31:53 crc kubenswrapper[4837]: I1001 10:31:53.788914 4837 scope.go:117] "RemoveContainer" containerID="17ae271dec3b8bd94a5ef7ea07bc0f7865f3f6661aa3ed242db6a45b1909a4ce" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.868180 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:29 crc kubenswrapper[4837]: E1001 10:32:29.884270 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cec3907b-97c0-47a7-b1c2-27ee5fd92148" containerName="collect-profiles" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.884715 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="cec3907b-97c0-47a7-b1c2-27ee5fd92148" containerName="collect-profiles" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.887729 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="cec3907b-97c0-47a7-b1c2-27ee5fd92148" containerName="collect-profiles" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.897545 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.960401 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.960473 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.960595 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9bf4\" (UniqueName: \"kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:29 crc kubenswrapper[4837]: I1001 10:32:29.983874 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.062848 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.062906 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.062967 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9bf4\" (UniqueName: \"kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.066021 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.066643 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.111299 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9bf4\" (UniqueName: \"kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4\") pod \"redhat-marketplace-vgg2h\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:30 crc kubenswrapper[4837]: I1001 10:32:30.225853 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:34 crc kubenswrapper[4837]: I1001 10:32:34.177687 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:34 crc kubenswrapper[4837]: W1001 10:32:34.231836 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2a0bd84_2139_4f27_acef_310473b47b14.slice/crio-fa1b270d801bbfc00897258f90f9d505bfe01ab0da0cb6d4c957cdaa3f6530c7 WatchSource:0}: Error finding container fa1b270d801bbfc00897258f90f9d505bfe01ab0da0cb6d4c957cdaa3f6530c7: Status 404 returned error can't find the container with id fa1b270d801bbfc00897258f90f9d505bfe01ab0da0cb6d4c957cdaa3f6530c7 Oct 01 10:32:34 crc kubenswrapper[4837]: I1001 10:32:34.293892 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerStarted","Data":"fa1b270d801bbfc00897258f90f9d505bfe01ab0da0cb6d4c957cdaa3f6530c7"} Oct 01 10:32:35 crc kubenswrapper[4837]: I1001 10:32:35.308005 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2a0bd84-2139-4f27-acef-310473b47b14" containerID="3eef38a5b2624a37a45ff18eab7fcb7c0e772a79be8db8f526bcd7ca4ba9711d" exitCode=0 Oct 01 10:32:35 crc kubenswrapper[4837]: I1001 10:32:35.308113 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerDied","Data":"3eef38a5b2624a37a45ff18eab7fcb7c0e772a79be8db8f526bcd7ca4ba9711d"} Oct 01 10:32:35 crc kubenswrapper[4837]: I1001 10:32:35.323023 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:32:37 crc kubenswrapper[4837]: I1001 10:32:37.332096 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2a0bd84-2139-4f27-acef-310473b47b14" containerID="a0e79aae0a8dbf36bca46d43af99e5a383d47e7ce3461b687c4ba6e804f12c5a" exitCode=0 Oct 01 10:32:37 crc kubenswrapper[4837]: I1001 10:32:37.332157 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerDied","Data":"a0e79aae0a8dbf36bca46d43af99e5a383d47e7ce3461b687c4ba6e804f12c5a"} Oct 01 10:32:38 crc kubenswrapper[4837]: I1001 10:32:38.349163 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerStarted","Data":"bfc2d56ffd3cfe1bac3a4c8c0686aa7855e7c392063ee59cdb7658b0f1ed826e"} Oct 01 10:32:38 crc kubenswrapper[4837]: I1001 10:32:38.373213 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vgg2h" podStartSLOduration=6.832301404 podStartE2EDuration="9.372781963s" podCreationTimestamp="2025-10-01 10:32:29 +0000 UTC" firstStartedPulling="2025-10-01 10:32:35.312968416 +0000 UTC m=+12412.154575861" lastFinishedPulling="2025-10-01 10:32:37.853448965 +0000 UTC m=+12414.695056420" observedRunningTime="2025-10-01 10:32:38.370709561 +0000 UTC m=+12415.212317036" watchObservedRunningTime="2025-10-01 10:32:38.372781963 +0000 UTC m=+12415.214389428" Oct 01 10:32:40 crc kubenswrapper[4837]: I1001 10:32:40.226547 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:40 crc kubenswrapper[4837]: I1001 10:32:40.226917 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:40 crc kubenswrapper[4837]: I1001 10:32:40.299266 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:50 crc kubenswrapper[4837]: I1001 10:32:50.289538 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:50 crc kubenswrapper[4837]: I1001 10:32:50.351318 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:50 crc kubenswrapper[4837]: I1001 10:32:50.519223 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vgg2h" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="registry-server" containerID="cri-o://bfc2d56ffd3cfe1bac3a4c8c0686aa7855e7c392063ee59cdb7658b0f1ed826e" gracePeriod=2 Oct 01 10:32:51 crc kubenswrapper[4837]: I1001 10:32:51.536912 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerDied","Data":"bfc2d56ffd3cfe1bac3a4c8c0686aa7855e7c392063ee59cdb7658b0f1ed826e"} Oct 01 10:32:51 crc kubenswrapper[4837]: I1001 10:32:51.536809 4837 generic.go:334] "Generic (PLEG): container finished" podID="d2a0bd84-2139-4f27-acef-310473b47b14" containerID="bfc2d56ffd3cfe1bac3a4c8c0686aa7855e7c392063ee59cdb7658b0f1ed826e" exitCode=0 Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.272292 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.370241 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9bf4\" (UniqueName: \"kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4\") pod \"d2a0bd84-2139-4f27-acef-310473b47b14\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.370518 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities\") pod \"d2a0bd84-2139-4f27-acef-310473b47b14\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.370606 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content\") pod \"d2a0bd84-2139-4f27-acef-310473b47b14\" (UID: \"d2a0bd84-2139-4f27-acef-310473b47b14\") " Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.372091 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities" (OuterVolumeSpecName: "utilities") pod "d2a0bd84-2139-4f27-acef-310473b47b14" (UID: "d2a0bd84-2139-4f27-acef-310473b47b14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.380360 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4" (OuterVolumeSpecName: "kube-api-access-d9bf4") pod "d2a0bd84-2139-4f27-acef-310473b47b14" (UID: "d2a0bd84-2139-4f27-acef-310473b47b14"). InnerVolumeSpecName "kube-api-access-d9bf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.388680 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2a0bd84-2139-4f27-acef-310473b47b14" (UID: "d2a0bd84-2139-4f27-acef-310473b47b14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.473990 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9bf4\" (UniqueName: \"kubernetes.io/projected/d2a0bd84-2139-4f27-acef-310473b47b14-kube-api-access-d9bf4\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.474025 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.474036 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2a0bd84-2139-4f27-acef-310473b47b14-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.551968 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgg2h" event={"ID":"d2a0bd84-2139-4f27-acef-310473b47b14","Type":"ContainerDied","Data":"fa1b270d801bbfc00897258f90f9d505bfe01ab0da0cb6d4c957cdaa3f6530c7"} Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.552038 4837 scope.go:117] "RemoveContainer" containerID="bfc2d56ffd3cfe1bac3a4c8c0686aa7855e7c392063ee59cdb7658b0f1ed826e" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.552089 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgg2h" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.585196 4837 scope.go:117] "RemoveContainer" containerID="a0e79aae0a8dbf36bca46d43af99e5a383d47e7ce3461b687c4ba6e804f12c5a" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.609424 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.624508 4837 scope.go:117] "RemoveContainer" containerID="3eef38a5b2624a37a45ff18eab7fcb7c0e772a79be8db8f526bcd7ca4ba9711d" Oct 01 10:32:52 crc kubenswrapper[4837]: I1001 10:32:52.631408 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgg2h"] Oct 01 10:32:53 crc kubenswrapper[4837]: I1001 10:32:53.851222 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" path="/var/lib/kubelet/pods/d2a0bd84-2139-4f27-acef-310473b47b14/volumes" Oct 01 10:33:53 crc kubenswrapper[4837]: I1001 10:33:53.083388 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:33:53 crc kubenswrapper[4837]: I1001 10:33:53.083910 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.796212 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:34:13 crc kubenswrapper[4837]: E1001 10:34:13.797014 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="extract-utilities" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.797025 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="extract-utilities" Oct 01 10:34:13 crc kubenswrapper[4837]: E1001 10:34:13.797052 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="extract-content" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.797058 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="extract-content" Oct 01 10:34:13 crc kubenswrapper[4837]: E1001 10:34:13.797084 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="registry-server" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.797091 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="registry-server" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.797335 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a0bd84-2139-4f27-acef-310473b47b14" containerName="registry-server" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.798782 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.839336 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.877413 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.878300 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.878684 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqjjf\" (UniqueName: \"kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.981122 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.981217 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.981248 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqjjf\" (UniqueName: \"kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.981793 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:13 crc kubenswrapper[4837]: I1001 10:34:13.982330 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:14 crc kubenswrapper[4837]: I1001 10:34:14.006147 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqjjf\" (UniqueName: \"kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf\") pod \"redhat-operators-5j26m\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:14 crc kubenswrapper[4837]: I1001 10:34:14.123177 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:15 crc kubenswrapper[4837]: I1001 10:34:15.146864 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:34:15 crc kubenswrapper[4837]: I1001 10:34:15.487156 4837 generic.go:334] "Generic (PLEG): container finished" podID="10a9fad5-13d8-46ec-82c5-227953c56557" containerID="4d9db6b1a40cd6cd31844f9c7eaeb43aa5ec6375aedbf055a3cfc11e55065a56" exitCode=0 Oct 01 10:34:15 crc kubenswrapper[4837]: I1001 10:34:15.487259 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerDied","Data":"4d9db6b1a40cd6cd31844f9c7eaeb43aa5ec6375aedbf055a3cfc11e55065a56"} Oct 01 10:34:15 crc kubenswrapper[4837]: I1001 10:34:15.487550 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerStarted","Data":"ab6f6e9b9263aa78dd56e5d732554f6c0f4c74e42777dcb547436766749d28b8"} Oct 01 10:34:17 crc kubenswrapper[4837]: I1001 10:34:17.509369 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerStarted","Data":"42272e1a5de564256ed2877f20ccf18cd72d0045b00eba44828371bdaee200a8"} Oct 01 10:34:22 crc kubenswrapper[4837]: I1001 10:34:22.562634 4837 generic.go:334] "Generic (PLEG): container finished" podID="10a9fad5-13d8-46ec-82c5-227953c56557" containerID="42272e1a5de564256ed2877f20ccf18cd72d0045b00eba44828371bdaee200a8" exitCode=0 Oct 01 10:34:22 crc kubenswrapper[4837]: I1001 10:34:22.562731 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerDied","Data":"42272e1a5de564256ed2877f20ccf18cd72d0045b00eba44828371bdaee200a8"} Oct 01 10:34:23 crc kubenswrapper[4837]: I1001 10:34:23.080076 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:34:23 crc kubenswrapper[4837]: I1001 10:34:23.080476 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:34:23 crc kubenswrapper[4837]: I1001 10:34:23.584653 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerStarted","Data":"5545a4fa339c596373021de59e5e2d34950fd4736b0eda2aaf8224f4c8079157"} Oct 01 10:34:23 crc kubenswrapper[4837]: I1001 10:34:23.620229 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5j26m" podStartSLOduration=3.071531516 podStartE2EDuration="10.620202895s" podCreationTimestamp="2025-10-01 10:34:13 +0000 UTC" firstStartedPulling="2025-10-01 10:34:15.491349392 +0000 UTC m=+12512.332956847" lastFinishedPulling="2025-10-01 10:34:23.040020771 +0000 UTC m=+12519.881628226" observedRunningTime="2025-10-01 10:34:23.604295821 +0000 UTC m=+12520.445903276" watchObservedRunningTime="2025-10-01 10:34:23.620202895 +0000 UTC m=+12520.461810350" Oct 01 10:34:24 crc kubenswrapper[4837]: I1001 10:34:24.123608 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:24 crc kubenswrapper[4837]: I1001 10:34:24.124034 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:34:25 crc kubenswrapper[4837]: I1001 10:34:25.180225 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" probeResult="failure" output=< Oct 01 10:34:25 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:34:25 crc kubenswrapper[4837]: > Oct 01 10:34:35 crc kubenswrapper[4837]: I1001 10:34:35.181865 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" probeResult="failure" output=< Oct 01 10:34:35 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:34:35 crc kubenswrapper[4837]: > Oct 01 10:34:45 crc kubenswrapper[4837]: I1001 10:34:45.175847 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" probeResult="failure" output=< Oct 01 10:34:45 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:34:45 crc kubenswrapper[4837]: > Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.079145 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.079713 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.079771 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.080556 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.080603 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" gracePeriod=600 Oct 01 10:34:53 crc kubenswrapper[4837]: E1001 10:34:53.242269 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.910639 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" exitCode=0 Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.910788 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821"} Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.911074 4837 scope.go:117] "RemoveContainer" containerID="d76fe06fc628281f8c62e9f8cf01d1bee7d450b18411a760edf02433e2f5c504" Oct 01 10:34:53 crc kubenswrapper[4837]: I1001 10:34:53.911966 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:34:53 crc kubenswrapper[4837]: E1001 10:34:53.912412 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:34:55 crc kubenswrapper[4837]: I1001 10:34:55.175121 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" probeResult="failure" output=< Oct 01 10:34:55 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:34:55 crc kubenswrapper[4837]: > Oct 01 10:35:05 crc kubenswrapper[4837]: I1001 10:35:05.189495 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" probeResult="failure" output=< Oct 01 10:35:05 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:35:05 crc kubenswrapper[4837]: > Oct 01 10:35:08 crc kubenswrapper[4837]: I1001 10:35:08.816762 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:35:08 crc kubenswrapper[4837]: E1001 10:35:08.817678 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:35:14 crc kubenswrapper[4837]: I1001 10:35:14.198152 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:35:14 crc kubenswrapper[4837]: I1001 10:35:14.256116 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:35:15 crc kubenswrapper[4837]: I1001 10:35:15.017883 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:35:16 crc kubenswrapper[4837]: I1001 10:35:16.213836 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5j26m" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" containerID="cri-o://5545a4fa339c596373021de59e5e2d34950fd4736b0eda2aaf8224f4c8079157" gracePeriod=2 Oct 01 10:35:17 crc kubenswrapper[4837]: I1001 10:35:17.242268 4837 generic.go:334] "Generic (PLEG): container finished" podID="10a9fad5-13d8-46ec-82c5-227953c56557" containerID="5545a4fa339c596373021de59e5e2d34950fd4736b0eda2aaf8224f4c8079157" exitCode=0 Oct 01 10:35:17 crc kubenswrapper[4837]: I1001 10:35:17.242622 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerDied","Data":"5545a4fa339c596373021de59e5e2d34950fd4736b0eda2aaf8224f4c8079157"} Oct 01 10:35:17 crc kubenswrapper[4837]: I1001 10:35:17.992605 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.023045 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content\") pod \"10a9fad5-13d8-46ec-82c5-227953c56557\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.023084 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqjjf\" (UniqueName: \"kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf\") pod \"10a9fad5-13d8-46ec-82c5-227953c56557\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.023192 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities\") pod \"10a9fad5-13d8-46ec-82c5-227953c56557\" (UID: \"10a9fad5-13d8-46ec-82c5-227953c56557\") " Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.024253 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities" (OuterVolumeSpecName: "utilities") pod "10a9fad5-13d8-46ec-82c5-227953c56557" (UID: "10a9fad5-13d8-46ec-82c5-227953c56557"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.051944 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf" (OuterVolumeSpecName: "kube-api-access-sqjjf") pod "10a9fad5-13d8-46ec-82c5-227953c56557" (UID: "10a9fad5-13d8-46ec-82c5-227953c56557"). InnerVolumeSpecName "kube-api-access-sqjjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.116661 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10a9fad5-13d8-46ec-82c5-227953c56557" (UID: "10a9fad5-13d8-46ec-82c5-227953c56557"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.125964 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.125996 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10a9fad5-13d8-46ec-82c5-227953c56557-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.126008 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqjjf\" (UniqueName: \"kubernetes.io/projected/10a9fad5-13d8-46ec-82c5-227953c56557-kube-api-access-sqjjf\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.254716 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5j26m" event={"ID":"10a9fad5-13d8-46ec-82c5-227953c56557","Type":"ContainerDied","Data":"ab6f6e9b9263aa78dd56e5d732554f6c0f4c74e42777dcb547436766749d28b8"} Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.254768 4837 scope.go:117] "RemoveContainer" containerID="5545a4fa339c596373021de59e5e2d34950fd4736b0eda2aaf8224f4c8079157" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.254895 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5j26m" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.294857 4837 scope.go:117] "RemoveContainer" containerID="42272e1a5de564256ed2877f20ccf18cd72d0045b00eba44828371bdaee200a8" Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.295627 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.308521 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5j26m"] Oct 01 10:35:18 crc kubenswrapper[4837]: I1001 10:35:18.339005 4837 scope.go:117] "RemoveContainer" containerID="4d9db6b1a40cd6cd31844f9c7eaeb43aa5ec6375aedbf055a3cfc11e55065a56" Oct 01 10:35:19 crc kubenswrapper[4837]: I1001 10:35:19.829412 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" path="/var/lib/kubelet/pods/10a9fad5-13d8-46ec-82c5-227953c56557/volumes" Oct 01 10:35:21 crc kubenswrapper[4837]: I1001 10:35:21.816565 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:35:21 crc kubenswrapper[4837]: E1001 10:35:21.817257 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:35:35 crc kubenswrapper[4837]: I1001 10:35:35.818205 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:35:35 crc kubenswrapper[4837]: E1001 10:35:35.818926 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:35:44 crc kubenswrapper[4837]: E1001 10:35:44.619682 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:35:47 crc kubenswrapper[4837]: I1001 10:35:47.816954 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:35:47 crc kubenswrapper[4837]: E1001 10:35:47.817600 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:35:54 crc kubenswrapper[4837]: E1001 10:35:54.927454 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:36:02 crc kubenswrapper[4837]: I1001 10:36:02.816299 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:36:02 crc kubenswrapper[4837]: E1001 10:36:02.816988 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:36:05 crc kubenswrapper[4837]: E1001 10:36:05.260593 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:36:15 crc kubenswrapper[4837]: E1001 10:36:15.531723 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:36:16 crc kubenswrapper[4837]: I1001 10:36:16.815684 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:36:16 crc kubenswrapper[4837]: E1001 10:36:16.816243 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:36:25 crc kubenswrapper[4837]: E1001 10:36:25.838063 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:36:27 crc kubenswrapper[4837]: I1001 10:36:27.816532 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:36:27 crc kubenswrapper[4837]: E1001 10:36:27.817227 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:36:36 crc kubenswrapper[4837]: E1001 10:36:36.126245 4837 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 01 10:36:40 crc kubenswrapper[4837]: I1001 10:36:40.816641 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:36:40 crc kubenswrapper[4837]: E1001 10:36:40.817304 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:36:51 crc kubenswrapper[4837]: I1001 10:36:51.815986 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:36:51 crc kubenswrapper[4837]: E1001 10:36:51.816632 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:37:06 crc kubenswrapper[4837]: I1001 10:37:06.815904 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:37:06 crc kubenswrapper[4837]: E1001 10:37:06.817092 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:37:17 crc kubenswrapper[4837]: I1001 10:37:17.816838 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:37:17 crc kubenswrapper[4837]: E1001 10:37:17.817616 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:37:32 crc kubenswrapper[4837]: I1001 10:37:32.816161 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:37:32 crc kubenswrapper[4837]: E1001 10:37:32.818033 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:37:44 crc kubenswrapper[4837]: I1001 10:37:44.815874 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:37:44 crc kubenswrapper[4837]: E1001 10:37:44.816757 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:37:58 crc kubenswrapper[4837]: I1001 10:37:58.816426 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:37:58 crc kubenswrapper[4837]: E1001 10:37:58.817300 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:38:10 crc kubenswrapper[4837]: I1001 10:38:10.816087 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:38:10 crc kubenswrapper[4837]: E1001 10:38:10.816800 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:38:25 crc kubenswrapper[4837]: I1001 10:38:25.817392 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:38:25 crc kubenswrapper[4837]: E1001 10:38:25.818290 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:38:37 crc kubenswrapper[4837]: I1001 10:38:37.816741 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:38:37 crc kubenswrapper[4837]: E1001 10:38:37.817538 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.585636 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:40 crc kubenswrapper[4837]: E1001 10:38:40.586783 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="extract-utilities" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.586804 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="extract-utilities" Oct 01 10:38:40 crc kubenswrapper[4837]: E1001 10:38:40.586827 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="extract-content" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.586835 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="extract-content" Oct 01 10:38:40 crc kubenswrapper[4837]: E1001 10:38:40.586856 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.586988 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.587207 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a9fad5-13d8-46ec-82c5-227953c56557" containerName="registry-server" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.589409 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.603625 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.733902 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.734020 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v4rs\" (UniqueName: \"kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.734085 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.836343 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.836511 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v4rs\" (UniqueName: \"kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.836906 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.836954 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.837370 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.860728 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v4rs\" (UniqueName: \"kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs\") pod \"community-operators-w8prg\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:40 crc kubenswrapper[4837]: I1001 10:38:40.921114 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:41 crc kubenswrapper[4837]: I1001 10:38:41.933922 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:42 crc kubenswrapper[4837]: I1001 10:38:42.484335 4837 generic.go:334] "Generic (PLEG): container finished" podID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerID="213a040eb51863a11becff5112b171315b8ee4655c0829cb5bde63b446e89150" exitCode=0 Oct 01 10:38:42 crc kubenswrapper[4837]: I1001 10:38:42.484394 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerDied","Data":"213a040eb51863a11becff5112b171315b8ee4655c0829cb5bde63b446e89150"} Oct 01 10:38:42 crc kubenswrapper[4837]: I1001 10:38:42.484419 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerStarted","Data":"991228fedec360de1238b927afff8895c75a9018c3efbf7bae44335b40d9dc55"} Oct 01 10:38:42 crc kubenswrapper[4837]: I1001 10:38:42.487787 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:38:44 crc kubenswrapper[4837]: I1001 10:38:44.507325 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerStarted","Data":"f84a5b54d3f50cf3484eb65a220de89994a06992958b4c4da6f743773c127c99"} Oct 01 10:38:45 crc kubenswrapper[4837]: I1001 10:38:45.518263 4837 generic.go:334] "Generic (PLEG): container finished" podID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerID="f84a5b54d3f50cf3484eb65a220de89994a06992958b4c4da6f743773c127c99" exitCode=0 Oct 01 10:38:45 crc kubenswrapper[4837]: I1001 10:38:45.518299 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerDied","Data":"f84a5b54d3f50cf3484eb65a220de89994a06992958b4c4da6f743773c127c99"} Oct 01 10:38:46 crc kubenswrapper[4837]: I1001 10:38:46.531118 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerStarted","Data":"7c65c04b3b9ecb375600272c5145c1f56a2f0d4b7d79b58f4ddbb8f38a056cf9"} Oct 01 10:38:46 crc kubenswrapper[4837]: I1001 10:38:46.558363 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w8prg" podStartSLOduration=2.847771474 podStartE2EDuration="6.558337464s" podCreationTimestamp="2025-10-01 10:38:40 +0000 UTC" firstStartedPulling="2025-10-01 10:38:42.487470113 +0000 UTC m=+12779.329077588" lastFinishedPulling="2025-10-01 10:38:46.198036123 +0000 UTC m=+12783.039643578" observedRunningTime="2025-10-01 10:38:46.549006064 +0000 UTC m=+12783.390613519" watchObservedRunningTime="2025-10-01 10:38:46.558337464 +0000 UTC m=+12783.399944929" Oct 01 10:38:49 crc kubenswrapper[4837]: I1001 10:38:49.816911 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:38:49 crc kubenswrapper[4837]: E1001 10:38:49.817874 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:38:50 crc kubenswrapper[4837]: I1001 10:38:50.922218 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:50 crc kubenswrapper[4837]: I1001 10:38:50.922294 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:50 crc kubenswrapper[4837]: I1001 10:38:50.996991 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:51 crc kubenswrapper[4837]: I1001 10:38:51.629502 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:51 crc kubenswrapper[4837]: I1001 10:38:51.680546 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:53 crc kubenswrapper[4837]: I1001 10:38:53.601416 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w8prg" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="registry-server" containerID="cri-o://7c65c04b3b9ecb375600272c5145c1f56a2f0d4b7d79b58f4ddbb8f38a056cf9" gracePeriod=2 Oct 01 10:38:54 crc kubenswrapper[4837]: I1001 10:38:54.617570 4837 generic.go:334] "Generic (PLEG): container finished" podID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerID="7c65c04b3b9ecb375600272c5145c1f56a2f0d4b7d79b58f4ddbb8f38a056cf9" exitCode=0 Oct 01 10:38:54 crc kubenswrapper[4837]: I1001 10:38:54.617635 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerDied","Data":"7c65c04b3b9ecb375600272c5145c1f56a2f0d4b7d79b58f4ddbb8f38a056cf9"} Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.276992 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.353444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content\") pod \"dfb166c9-2c0d-427f-90d9-b98e744accb6\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.353533 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities\") pod \"dfb166c9-2c0d-427f-90d9-b98e744accb6\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.353674 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v4rs\" (UniqueName: \"kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs\") pod \"dfb166c9-2c0d-427f-90d9-b98e744accb6\" (UID: \"dfb166c9-2c0d-427f-90d9-b98e744accb6\") " Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.357245 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities" (OuterVolumeSpecName: "utilities") pod "dfb166c9-2c0d-427f-90d9-b98e744accb6" (UID: "dfb166c9-2c0d-427f-90d9-b98e744accb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.384419 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs" (OuterVolumeSpecName: "kube-api-access-2v4rs") pod "dfb166c9-2c0d-427f-90d9-b98e744accb6" (UID: "dfb166c9-2c0d-427f-90d9-b98e744accb6"). InnerVolumeSpecName "kube-api-access-2v4rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.434701 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfb166c9-2c0d-427f-90d9-b98e744accb6" (UID: "dfb166c9-2c0d-427f-90d9-b98e744accb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.456049 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.456090 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v4rs\" (UniqueName: \"kubernetes.io/projected/dfb166c9-2c0d-427f-90d9-b98e744accb6-kube-api-access-2v4rs\") on node \"crc\" DevicePath \"\"" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.456101 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb166c9-2c0d-427f-90d9-b98e744accb6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.632633 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8prg" event={"ID":"dfb166c9-2c0d-427f-90d9-b98e744accb6","Type":"ContainerDied","Data":"991228fedec360de1238b927afff8895c75a9018c3efbf7bae44335b40d9dc55"} Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.632710 4837 scope.go:117] "RemoveContainer" containerID="7c65c04b3b9ecb375600272c5145c1f56a2f0d4b7d79b58f4ddbb8f38a056cf9" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.632904 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8prg" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.667198 4837 scope.go:117] "RemoveContainer" containerID="f84a5b54d3f50cf3484eb65a220de89994a06992958b4c4da6f743773c127c99" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.670347 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.682406 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w8prg"] Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.694658 4837 scope.go:117] "RemoveContainer" containerID="213a040eb51863a11becff5112b171315b8ee4655c0829cb5bde63b446e89150" Oct 01 10:38:55 crc kubenswrapper[4837]: I1001 10:38:55.827585 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" path="/var/lib/kubelet/pods/dfb166c9-2c0d-427f-90d9-b98e744accb6/volumes" Oct 01 10:39:04 crc kubenswrapper[4837]: I1001 10:39:04.816561 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:39:04 crc kubenswrapper[4837]: E1001 10:39:04.817459 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:39:19 crc kubenswrapper[4837]: I1001 10:39:19.815901 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:39:19 crc kubenswrapper[4837]: E1001 10:39:19.816729 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:39:30 crc kubenswrapper[4837]: I1001 10:39:30.816367 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:39:30 crc kubenswrapper[4837]: E1001 10:39:30.820870 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:39:41 crc kubenswrapper[4837]: I1001 10:39:41.816239 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:39:41 crc kubenswrapper[4837]: E1001 10:39:41.816964 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:39:53 crc kubenswrapper[4837]: I1001 10:39:53.823126 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:39:54 crc kubenswrapper[4837]: I1001 10:39:54.217525 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304"} Oct 01 10:40:25 crc kubenswrapper[4837]: I1001 10:40:25.605801 4837 generic.go:334] "Generic (PLEG): container finished" podID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" containerID="047cf8eb1f04848fdf1eb08f432a6af8d7c90f7ec3ff78c5774c2381d196b016" exitCode=0 Oct 01 10:40:25 crc kubenswrapper[4837]: I1001 10:40:25.606142 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"62fa4c75-1944-4f6f-8efe-8463b065c5c5","Type":"ContainerDied","Data":"047cf8eb1f04848fdf1eb08f432a6af8d7c90f7ec3ff78c5774c2381d196b016"} Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.214867 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348399 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz92l\" (UniqueName: \"kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348475 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348504 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348567 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348608 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348676 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348776 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348824 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.348901 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\" (UID: \"62fa4c75-1944-4f6f-8efe-8463b065c5c5\") " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.349748 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.349764 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data" (OuterVolumeSpecName: "config-data") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.350324 4837 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.350353 4837 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.353428 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.359919 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.374939 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l" (OuterVolumeSpecName: "kube-api-access-gz92l") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "kube-api-access-gz92l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.400010 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.415814 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.420066 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.433502 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "62fa4c75-1944-4f6f-8efe-8463b065c5c5" (UID: "62fa4c75-1944-4f6f-8efe-8463b065c5c5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.451953 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452147 4837 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/62fa4c75-1944-4f6f-8efe-8463b065c5c5-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452216 4837 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452330 4837 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452410 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz92l\" (UniqueName: \"kubernetes.io/projected/62fa4c75-1944-4f6f-8efe-8463b065c5c5-kube-api-access-gz92l\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452492 4837 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/62fa4c75-1944-4f6f-8efe-8463b065c5c5-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.452566 4837 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/62fa4c75-1944-4f6f-8efe-8463b065c5c5-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.481514 4837 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.554582 4837 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.635943 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"62fa4c75-1944-4f6f-8efe-8463b065c5c5","Type":"ContainerDied","Data":"e0e51fd56d35bf27d3c4af2d6e5f58ce916d11bf307097afa6316dd74b85ff7a"} Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.635991 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0e51fd56d35bf27d3c4af2d6e5f58ce916d11bf307097afa6316dd74b85ff7a" Oct 01 10:40:28 crc kubenswrapper[4837]: I1001 10:40:28.636005 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.217316 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:40:38 crc kubenswrapper[4837]: E1001 10:40:38.218227 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="extract-content" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218244 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="extract-content" Oct 01 10:40:38 crc kubenswrapper[4837]: E1001 10:40:38.218269 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218293 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:40:38 crc kubenswrapper[4837]: E1001 10:40:38.218310 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="registry-server" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218318 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="registry-server" Oct 01 10:40:38 crc kubenswrapper[4837]: E1001 10:40:38.218359 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="extract-utilities" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218369 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="extract-utilities" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218670 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fa4c75-1944-4f6f-8efe-8463b065c5c5" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.218706 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb166c9-2c0d-427f-90d9-b98e744accb6" containerName="registry-server" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.219602 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.229059 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.239730 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cz4sn" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.273855 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rstt7\" (UniqueName: \"kubernetes.io/projected/c57e86bb-3dca-4c86-938a-4072fa900d0c-kube-api-access-rstt7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.274043 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.376302 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rstt7\" (UniqueName: \"kubernetes.io/projected/c57e86bb-3dca-4c86-938a-4072fa900d0c-kube-api-access-rstt7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.376586 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.378512 4837 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.408014 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rstt7\" (UniqueName: \"kubernetes.io/projected/c57e86bb-3dca-4c86-938a-4072fa900d0c-kube-api-access-rstt7\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.411847 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c57e86bb-3dca-4c86-938a-4072fa900d0c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:38 crc kubenswrapper[4837]: I1001 10:40:38.555087 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:40:39 crc kubenswrapper[4837]: I1001 10:40:39.367658 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:40:39 crc kubenswrapper[4837]: W1001 10:40:39.369789 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc57e86bb_3dca_4c86_938a_4072fa900d0c.slice/crio-ba0934799efdcbb8dff7741161c5e388decdedf5b5c5b203c6b9c76c13df250e WatchSource:0}: Error finding container ba0934799efdcbb8dff7741161c5e388decdedf5b5c5b203c6b9c76c13df250e: Status 404 returned error can't find the container with id ba0934799efdcbb8dff7741161c5e388decdedf5b5c5b203c6b9c76c13df250e Oct 01 10:40:39 crc kubenswrapper[4837]: I1001 10:40:39.744222 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c57e86bb-3dca-4c86-938a-4072fa900d0c","Type":"ContainerStarted","Data":"ba0934799efdcbb8dff7741161c5e388decdedf5b5c5b203c6b9c76c13df250e"} Oct 01 10:40:41 crc kubenswrapper[4837]: I1001 10:40:41.767968 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c57e86bb-3dca-4c86-938a-4072fa900d0c","Type":"ContainerStarted","Data":"4660240d89009d213ba6e8150dd4f87e46aa2b53349c1aa6eee87bedd3335d79"} Oct 01 10:40:41 crc kubenswrapper[4837]: I1001 10:40:41.789805 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.4369035390000002 podStartE2EDuration="3.789715805s" podCreationTimestamp="2025-10-01 10:40:38 +0000 UTC" firstStartedPulling="2025-10-01 10:40:39.372982085 +0000 UTC m=+12896.214589540" lastFinishedPulling="2025-10-01 10:40:40.725794351 +0000 UTC m=+12897.567401806" observedRunningTime="2025-10-01 10:40:41.780907037 +0000 UTC m=+12898.622514522" watchObservedRunningTime="2025-10-01 10:40:41.789715805 +0000 UTC m=+12898.631323270" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.588589 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.592044 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.600404 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.781471 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8cjz\" (UniqueName: \"kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.782254 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.782366 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.884822 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8cjz\" (UniqueName: \"kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.884995 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.885082 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.885584 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.885922 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.922071 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8cjz\" (UniqueName: \"kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz\") pod \"certified-operators-dtp6f\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:32 crc kubenswrapper[4837]: I1001 10:41:32.925063 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:33 crc kubenswrapper[4837]: I1001 10:41:33.856614 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:33 crc kubenswrapper[4837]: W1001 10:41:33.858893 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod583f05e2_b5ac_4f0c_a1a4_de7cbe32d662.slice/crio-f40b259806d32c0cbc7320bb9e4e787df8c530386e9cbb97cf2d4cbe40c0384e WatchSource:0}: Error finding container f40b259806d32c0cbc7320bb9e4e787df8c530386e9cbb97cf2d4cbe40c0384e: Status 404 returned error can't find the container with id f40b259806d32c0cbc7320bb9e4e787df8c530386e9cbb97cf2d4cbe40c0384e Oct 01 10:41:34 crc kubenswrapper[4837]: I1001 10:41:34.323369 4837 generic.go:334] "Generic (PLEG): container finished" podID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerID="981745543d9a06d3c2ab822ff3e8bca14ba7cff7d38b4282780c48c84b16b9ab" exitCode=0 Oct 01 10:41:34 crc kubenswrapper[4837]: I1001 10:41:34.323446 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerDied","Data":"981745543d9a06d3c2ab822ff3e8bca14ba7cff7d38b4282780c48c84b16b9ab"} Oct 01 10:41:34 crc kubenswrapper[4837]: I1001 10:41:34.323673 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerStarted","Data":"f40b259806d32c0cbc7320bb9e4e787df8c530386e9cbb97cf2d4cbe40c0384e"} Oct 01 10:41:36 crc kubenswrapper[4837]: I1001 10:41:36.342738 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerStarted","Data":"9f870fc955a692576fc2cf2da0905f4f5ce56baeb0d44967f3fe2721306ba3d2"} Oct 01 10:41:37 crc kubenswrapper[4837]: I1001 10:41:37.353357 4837 generic.go:334] "Generic (PLEG): container finished" podID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerID="9f870fc955a692576fc2cf2da0905f4f5ce56baeb0d44967f3fe2721306ba3d2" exitCode=0 Oct 01 10:41:37 crc kubenswrapper[4837]: I1001 10:41:37.353458 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerDied","Data":"9f870fc955a692576fc2cf2da0905f4f5ce56baeb0d44967f3fe2721306ba3d2"} Oct 01 10:41:38 crc kubenswrapper[4837]: I1001 10:41:38.364972 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerStarted","Data":"ca319ee9078620248f8abd50f6a31d4bfbed16700599547bbb69c0f938e0b8aa"} Oct 01 10:41:38 crc kubenswrapper[4837]: I1001 10:41:38.389843 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtp6f" podStartSLOduration=2.959336323 podStartE2EDuration="6.389821169s" podCreationTimestamp="2025-10-01 10:41:32 +0000 UTC" firstStartedPulling="2025-10-01 10:41:34.325542161 +0000 UTC m=+12951.167149636" lastFinishedPulling="2025-10-01 10:41:37.756027027 +0000 UTC m=+12954.597634482" observedRunningTime="2025-10-01 10:41:38.379842443 +0000 UTC m=+12955.221449908" watchObservedRunningTime="2025-10-01 10:41:38.389821169 +0000 UTC m=+12955.231428624" Oct 01 10:41:42 crc kubenswrapper[4837]: I1001 10:41:42.925906 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:42 crc kubenswrapper[4837]: I1001 10:41:42.926273 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:42 crc kubenswrapper[4837]: I1001 10:41:42.972177 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:43 crc kubenswrapper[4837]: I1001 10:41:43.474048 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:43 crc kubenswrapper[4837]: I1001 10:41:43.519908 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:45 crc kubenswrapper[4837]: I1001 10:41:45.446815 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtp6f" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="registry-server" containerID="cri-o://ca319ee9078620248f8abd50f6a31d4bfbed16700599547bbb69c0f938e0b8aa" gracePeriod=2 Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.463144 4837 generic.go:334] "Generic (PLEG): container finished" podID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerID="ca319ee9078620248f8abd50f6a31d4bfbed16700599547bbb69c0f938e0b8aa" exitCode=0 Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.463574 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerDied","Data":"ca319ee9078620248f8abd50f6a31d4bfbed16700599547bbb69c0f938e0b8aa"} Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.570556 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.679885 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8cjz\" (UniqueName: \"kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz\") pod \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.680327 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content\") pod \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.680449 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities\") pod \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\" (UID: \"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662\") " Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.681772 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities" (OuterVolumeSpecName: "utilities") pod "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" (UID: "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.692095 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz" (OuterVolumeSpecName: "kube-api-access-v8cjz") pod "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" (UID: "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662"). InnerVolumeSpecName "kube-api-access-v8cjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.728475 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" (UID: "583f05e2-b5ac-4f0c-a1a4-de7cbe32d662"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.782978 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.783011 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:41:46 crc kubenswrapper[4837]: I1001 10:41:46.783022 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8cjz\" (UniqueName: \"kubernetes.io/projected/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662-kube-api-access-v8cjz\") on node \"crc\" DevicePath \"\"" Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.490325 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtp6f" event={"ID":"583f05e2-b5ac-4f0c-a1a4-de7cbe32d662","Type":"ContainerDied","Data":"f40b259806d32c0cbc7320bb9e4e787df8c530386e9cbb97cf2d4cbe40c0384e"} Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.490398 4837 scope.go:117] "RemoveContainer" containerID="ca319ee9078620248f8abd50f6a31d4bfbed16700599547bbb69c0f938e0b8aa" Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.490656 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtp6f" Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.513149 4837 scope.go:117] "RemoveContainer" containerID="9f870fc955a692576fc2cf2da0905f4f5ce56baeb0d44967f3fe2721306ba3d2" Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.539648 4837 scope.go:117] "RemoveContainer" containerID="981745543d9a06d3c2ab822ff3e8bca14ba7cff7d38b4282780c48c84b16b9ab" Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.544559 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.562528 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtp6f"] Oct 01 10:41:47 crc kubenswrapper[4837]: I1001 10:41:47.834438 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" path="/var/lib/kubelet/pods/583f05e2-b5ac-4f0c-a1a4-de7cbe32d662/volumes" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.507144 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ntfc2/must-gather-6rdm2"] Oct 01 10:41:57 crc kubenswrapper[4837]: E1001 10:41:57.508037 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="extract-utilities" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.508049 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="extract-utilities" Oct 01 10:41:57 crc kubenswrapper[4837]: E1001 10:41:57.508069 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="registry-server" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.508075 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="registry-server" Oct 01 10:41:57 crc kubenswrapper[4837]: E1001 10:41:57.508097 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="extract-content" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.508103 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="extract-content" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.508330 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="583f05e2-b5ac-4f0c-a1a4-de7cbe32d662" containerName="registry-server" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.509426 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.513087 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ntfc2"/"default-dockercfg-fh8t5" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.514378 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ntfc2"/"kube-root-ca.crt" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.518033 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ntfc2"/"openshift-service-ca.crt" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.523776 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ntfc2/must-gather-6rdm2"] Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.604972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbm5\" (UniqueName: \"kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.605090 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.707547 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbm5\" (UniqueName: \"kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.707741 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.708132 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.737814 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbm5\" (UniqueName: \"kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5\") pod \"must-gather-6rdm2\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:57 crc kubenswrapper[4837]: I1001 10:41:57.830810 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:41:58 crc kubenswrapper[4837]: I1001 10:41:58.685616 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ntfc2/must-gather-6rdm2"] Oct 01 10:41:59 crc kubenswrapper[4837]: I1001 10:41:59.652038 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" event={"ID":"832b5f04-3558-43f1-84b3-abb7a96ba682","Type":"ContainerStarted","Data":"a2cabc83bf1f2faefbbc55e120a21d2d9afff366d64ed54019b725c2abf9330a"} Oct 01 10:42:03 crc kubenswrapper[4837]: I1001 10:42:03.715724 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" event={"ID":"832b5f04-3558-43f1-84b3-abb7a96ba682","Type":"ContainerStarted","Data":"ef8f14431fb346eb25686538db0ee7e102dd8a3b9f31148c0a0857c1012bf857"} Oct 01 10:42:04 crc kubenswrapper[4837]: I1001 10:42:04.727817 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" event={"ID":"832b5f04-3558-43f1-84b3-abb7a96ba682","Type":"ContainerStarted","Data":"ac02b73ec4edc289f196cc7ffd679bed047ad706f196286c6dc1072fff34e61a"} Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.732401 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" podStartSLOduration=8.43541971 podStartE2EDuration="12.732379364s" podCreationTimestamp="2025-10-01 10:41:57 +0000 UTC" firstStartedPulling="2025-10-01 10:41:58.690831847 +0000 UTC m=+12975.532439302" lastFinishedPulling="2025-10-01 10:42:02.987791501 +0000 UTC m=+12979.829398956" observedRunningTime="2025-10-01 10:42:04.745188488 +0000 UTC m=+12981.586795943" watchObservedRunningTime="2025-10-01 10:42:09.732379364 +0000 UTC m=+12986.573986819" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.741191 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-gpslx"] Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.746671 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.850487 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.850972 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bzzz\" (UniqueName: \"kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.952902 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bzzz\" (UniqueName: \"kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.953061 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.954001 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:09 crc kubenswrapper[4837]: I1001 10:42:09.971134 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bzzz\" (UniqueName: \"kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz\") pod \"crc-debug-gpslx\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:10 crc kubenswrapper[4837]: I1001 10:42:10.063862 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:42:10 crc kubenswrapper[4837]: I1001 10:42:10.795413 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" event={"ID":"6726f4a7-ef64-4db0-a5ad-4509d6b64302","Type":"ContainerStarted","Data":"5195f06ebaf683cd70857053a765d452251658bcdb197945601676c3e3eb109c"} Oct 01 10:42:23 crc kubenswrapper[4837]: I1001 10:42:23.080035 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:42:23 crc kubenswrapper[4837]: I1001 10:42:23.080548 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:42:23 crc kubenswrapper[4837]: I1001 10:42:23.948188 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" event={"ID":"6726f4a7-ef64-4db0-a5ad-4509d6b64302","Type":"ContainerStarted","Data":"5385e26cd1e83deb1712197807adbedf03697b0deeca3062274948b84b6d97d2"} Oct 01 10:42:23 crc kubenswrapper[4837]: I1001 10:42:23.976666 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" podStartSLOduration=1.8189760179999999 podStartE2EDuration="14.976647583s" podCreationTimestamp="2025-10-01 10:42:09 +0000 UTC" firstStartedPulling="2025-10-01 10:42:10.116507653 +0000 UTC m=+12986.958115108" lastFinishedPulling="2025-10-01 10:42:23.274179218 +0000 UTC m=+13000.115786673" observedRunningTime="2025-10-01 10:42:23.966596125 +0000 UTC m=+13000.808203580" watchObservedRunningTime="2025-10-01 10:42:23.976647583 +0000 UTC m=+13000.818255028" Oct 01 10:42:53 crc kubenswrapper[4837]: I1001 10:42:53.079304 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:42:53 crc kubenswrapper[4837]: I1001 10:42:53.079853 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.079200 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.079930 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.080009 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.081118 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.081266 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304" gracePeriod=600 Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.597417 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304" exitCode=0 Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.597543 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304"} Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.597941 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa"} Oct 01 10:43:23 crc kubenswrapper[4837]: I1001 10:43:23.597972 4837 scope.go:117] "RemoveContainer" containerID="ce64c02ca487a7c326d99840f7968c500260a0dd5d43f74356fec27324e07821" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.229981 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq"] Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.233088 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.234781 4837 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.247736 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq"] Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.251151 4837 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.327179 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.327241 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crhqm\" (UniqueName: \"kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.327268 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.429415 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.429482 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crhqm\" (UniqueName: \"kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.429508 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.430214 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.447150 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.453297 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crhqm\" (UniqueName: \"kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm\") pod \"collect-profiles-29321925-5gmjq\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:00 crc kubenswrapper[4837]: I1001 10:45:00.559102 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:02 crc kubenswrapper[4837]: I1001 10:45:02.012444 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq"] Oct 01 10:45:02 crc kubenswrapper[4837]: I1001 10:45:02.642420 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" event={"ID":"85d7cf68-a329-489e-9941-39ff96fe0034","Type":"ContainerStarted","Data":"343155fb1ef8e56c16245cafb6bf9156b7749ff605ddcc8617684e5f19c63c22"} Oct 01 10:45:02 crc kubenswrapper[4837]: I1001 10:45:02.642671 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" event={"ID":"85d7cf68-a329-489e-9941-39ff96fe0034","Type":"ContainerStarted","Data":"3d352023d14fe84b895d4ce9477e12c005a6e15af3e733026b818172b4439eed"} Oct 01 10:45:02 crc kubenswrapper[4837]: I1001 10:45:02.657376 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" podStartSLOduration=2.657364324 podStartE2EDuration="2.657364324s" podCreationTimestamp="2025-10-01 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:45:02.655920878 +0000 UTC m=+13159.497528343" watchObservedRunningTime="2025-10-01 10:45:02.657364324 +0000 UTC m=+13159.498971779" Oct 01 10:45:04 crc kubenswrapper[4837]: I1001 10:45:04.662240 4837 generic.go:334] "Generic (PLEG): container finished" podID="85d7cf68-a329-489e-9941-39ff96fe0034" containerID="343155fb1ef8e56c16245cafb6bf9156b7749ff605ddcc8617684e5f19c63c22" exitCode=0 Oct 01 10:45:04 crc kubenswrapper[4837]: I1001 10:45:04.662319 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" event={"ID":"85d7cf68-a329-489e-9941-39ff96fe0034","Type":"ContainerDied","Data":"343155fb1ef8e56c16245cafb6bf9156b7749ff605ddcc8617684e5f19c63c22"} Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.341874 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.476387 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crhqm\" (UniqueName: \"kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm\") pod \"85d7cf68-a329-489e-9941-39ff96fe0034\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.476494 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume\") pod \"85d7cf68-a329-489e-9941-39ff96fe0034\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.476634 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume\") pod \"85d7cf68-a329-489e-9941-39ff96fe0034\" (UID: \"85d7cf68-a329-489e-9941-39ff96fe0034\") " Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.477365 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume" (OuterVolumeSpecName: "config-volume") pod "85d7cf68-a329-489e-9941-39ff96fe0034" (UID: "85d7cf68-a329-489e-9941-39ff96fe0034"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.485197 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "85d7cf68-a329-489e-9941-39ff96fe0034" (UID: "85d7cf68-a329-489e-9941-39ff96fe0034"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.500894 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm" (OuterVolumeSpecName: "kube-api-access-crhqm") pod "85d7cf68-a329-489e-9941-39ff96fe0034" (UID: "85d7cf68-a329-489e-9941-39ff96fe0034"). InnerVolumeSpecName "kube-api-access-crhqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.579004 4837 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/85d7cf68-a329-489e-9941-39ff96fe0034-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.579044 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crhqm\" (UniqueName: \"kubernetes.io/projected/85d7cf68-a329-489e-9941-39ff96fe0034-kube-api-access-crhqm\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.579053 4837 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/85d7cf68-a329-489e-9941-39ff96fe0034-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.691323 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" event={"ID":"85d7cf68-a329-489e-9941-39ff96fe0034","Type":"ContainerDied","Data":"3d352023d14fe84b895d4ce9477e12c005a6e15af3e733026b818172b4439eed"} Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.691364 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d352023d14fe84b895d4ce9477e12c005a6e15af3e733026b818172b4439eed" Oct 01 10:45:07 crc kubenswrapper[4837]: I1001 10:45:07.691372 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-5gmjq" Oct 01 10:45:08 crc kubenswrapper[4837]: I1001 10:45:08.478629 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl"] Oct 01 10:45:08 crc kubenswrapper[4837]: I1001 10:45:08.490749 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-6rsjl"] Oct 01 10:45:09 crc kubenswrapper[4837]: I1001 10:45:09.829777 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="796d73ac-53c6-4359-bdb9-a311ef6a597e" path="/var/lib/kubelet/pods/796d73ac-53c6-4359-bdb9-a311ef6a597e/volumes" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.445513 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:45:21 crc kubenswrapper[4837]: E1001 10:45:21.446519 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d7cf68-a329-489e-9941-39ff96fe0034" containerName="collect-profiles" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.446531 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d7cf68-a329-489e-9941-39ff96fe0034" containerName="collect-profiles" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.446798 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d7cf68-a329-489e-9941-39ff96fe0034" containerName="collect-profiles" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.448368 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.474907 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.584922 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vbcs\" (UniqueName: \"kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.585011 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.585029 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.688356 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vbcs\" (UniqueName: \"kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.688535 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.688570 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.689601 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.690139 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.713986 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vbcs\" (UniqueName: \"kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs\") pod \"redhat-operators-x5m4r\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:21 crc kubenswrapper[4837]: I1001 10:45:21.783747 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:22 crc kubenswrapper[4837]: I1001 10:45:22.514638 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:45:22 crc kubenswrapper[4837]: I1001 10:45:22.822092 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerID="9c3ad630b4d5b753f457fd115c6a3e069cbfc7c07f09643c0314046253e634b1" exitCode=0 Oct 01 10:45:22 crc kubenswrapper[4837]: I1001 10:45:22.822128 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerDied","Data":"9c3ad630b4d5b753f457fd115c6a3e069cbfc7c07f09643c0314046253e634b1"} Oct 01 10:45:22 crc kubenswrapper[4837]: I1001 10:45:22.822153 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerStarted","Data":"b75a433bc0b47c3232d05001538bfcb7932ccc4c089c401f27eac4da0ca98e26"} Oct 01 10:45:22 crc kubenswrapper[4837]: I1001 10:45:22.824068 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:45:23 crc kubenswrapper[4837]: I1001 10:45:23.081767 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:45:23 crc kubenswrapper[4837]: I1001 10:45:23.082030 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:45:24 crc kubenswrapper[4837]: I1001 10:45:24.850284 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerStarted","Data":"133655a0504063b9c995de189ff9a4e3e3d903e3971d68f3309c04c7e24fa749"} Oct 01 10:45:26 crc kubenswrapper[4837]: I1001 10:45:26.633468 4837 scope.go:117] "RemoveContainer" containerID="5a14feb6a805582b9e1a76e670633d8008329f567ec75e4257ff0ac5bca98bf8" Oct 01 10:45:28 crc kubenswrapper[4837]: I1001 10:45:28.891847 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerID="133655a0504063b9c995de189ff9a4e3e3d903e3971d68f3309c04c7e24fa749" exitCode=0 Oct 01 10:45:28 crc kubenswrapper[4837]: I1001 10:45:28.891893 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerDied","Data":"133655a0504063b9c995de189ff9a4e3e3d903e3971d68f3309c04c7e24fa749"} Oct 01 10:45:29 crc kubenswrapper[4837]: I1001 10:45:29.905532 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerStarted","Data":"2439747f810c4a0fc9a136337b338d25cf27821b26492a5ef34a1ede74f44010"} Oct 01 10:45:29 crc kubenswrapper[4837]: I1001 10:45:29.928849 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x5m4r" podStartSLOduration=2.434007902 podStartE2EDuration="8.92883134s" podCreationTimestamp="2025-10-01 10:45:21 +0000 UTC" firstStartedPulling="2025-10-01 10:45:22.823853093 +0000 UTC m=+13179.665460548" lastFinishedPulling="2025-10-01 10:45:29.318676541 +0000 UTC m=+13186.160283986" observedRunningTime="2025-10-01 10:45:29.921798286 +0000 UTC m=+13186.763405751" watchObservedRunningTime="2025-10-01 10:45:29.92883134 +0000 UTC m=+13186.770438795" Oct 01 10:45:31 crc kubenswrapper[4837]: I1001 10:45:31.784895 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:31 crc kubenswrapper[4837]: I1001 10:45:31.785164 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:45:32 crc kubenswrapper[4837]: I1001 10:45:32.849948 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5m4r" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" probeResult="failure" output=< Oct 01 10:45:32 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:45:32 crc kubenswrapper[4837]: > Oct 01 10:45:35 crc kubenswrapper[4837]: I1001 10:45:35.962004 4837 generic.go:334] "Generic (PLEG): container finished" podID="6726f4a7-ef64-4db0-a5ad-4509d6b64302" containerID="5385e26cd1e83deb1712197807adbedf03697b0deeca3062274948b84b6d97d2" exitCode=0 Oct 01 10:45:35 crc kubenswrapper[4837]: I1001 10:45:35.962079 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" event={"ID":"6726f4a7-ef64-4db0-a5ad-4509d6b64302","Type":"ContainerDied","Data":"5385e26cd1e83deb1712197807adbedf03697b0deeca3062274948b84b6d97d2"} Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.088298 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.120962 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-gpslx"] Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.130724 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-gpslx"] Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.228290 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bzzz\" (UniqueName: \"kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz\") pod \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.228436 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host\") pod \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\" (UID: \"6726f4a7-ef64-4db0-a5ad-4509d6b64302\") " Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.228788 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host" (OuterVolumeSpecName: "host") pod "6726f4a7-ef64-4db0-a5ad-4509d6b64302" (UID: "6726f4a7-ef64-4db0-a5ad-4509d6b64302"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.229439 4837 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6726f4a7-ef64-4db0-a5ad-4509d6b64302-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.234281 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz" (OuterVolumeSpecName: "kube-api-access-4bzzz") pod "6726f4a7-ef64-4db0-a5ad-4509d6b64302" (UID: "6726f4a7-ef64-4db0-a5ad-4509d6b64302"). InnerVolumeSpecName "kube-api-access-4bzzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.331108 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bzzz\" (UniqueName: \"kubernetes.io/projected/6726f4a7-ef64-4db0-a5ad-4509d6b64302-kube-api-access-4bzzz\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.827980 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6726f4a7-ef64-4db0-a5ad-4509d6b64302" path="/var/lib/kubelet/pods/6726f4a7-ef64-4db0-a5ad-4509d6b64302/volumes" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.982648 4837 scope.go:117] "RemoveContainer" containerID="5385e26cd1e83deb1712197807adbedf03697b0deeca3062274948b84b6d97d2" Oct 01 10:45:37 crc kubenswrapper[4837]: I1001 10:45:37.982746 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-gpslx" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.622780 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-ct5pr"] Oct 01 10:45:38 crc kubenswrapper[4837]: E1001 10:45:38.623192 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6726f4a7-ef64-4db0-a5ad-4509d6b64302" containerName="container-00" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.623206 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="6726f4a7-ef64-4db0-a5ad-4509d6b64302" containerName="container-00" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.623440 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="6726f4a7-ef64-4db0-a5ad-4509d6b64302" containerName="container-00" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.624160 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.766797 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww87n\" (UniqueName: \"kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.766957 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.869306 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww87n\" (UniqueName: \"kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.869830 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.869945 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.892896 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww87n\" (UniqueName: \"kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n\") pod \"crc-debug-ct5pr\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:38 crc kubenswrapper[4837]: I1001 10:45:38.948836 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:40 crc kubenswrapper[4837]: I1001 10:45:40.012306 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" event={"ID":"8985cb77-9aeb-4c86-83b2-c93d786c4a04","Type":"ContainerStarted","Data":"99289b4707c62fc27651295fbf0ad8dad630b570b04a6a3baf6fc36503a93c5a"} Oct 01 10:45:40 crc kubenswrapper[4837]: I1001 10:45:40.012570 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" event={"ID":"8985cb77-9aeb-4c86-83b2-c93d786c4a04","Type":"ContainerStarted","Data":"c2d3db9e62ddfad044c51a5e8d5109f8d65df45839a09acc1e7aac41cb58de57"} Oct 01 10:45:40 crc kubenswrapper[4837]: I1001 10:45:40.035124 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" podStartSLOduration=2.0351056339999998 podStartE2EDuration="2.035105634s" podCreationTimestamp="2025-10-01 10:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:45:40.024109944 +0000 UTC m=+13196.865717399" watchObservedRunningTime="2025-10-01 10:45:40.035105634 +0000 UTC m=+13196.876713079" Oct 01 10:45:41 crc kubenswrapper[4837]: I1001 10:45:41.043059 4837 generic.go:334] "Generic (PLEG): container finished" podID="8985cb77-9aeb-4c86-83b2-c93d786c4a04" containerID="99289b4707c62fc27651295fbf0ad8dad630b570b04a6a3baf6fc36503a93c5a" exitCode=0 Oct 01 10:45:41 crc kubenswrapper[4837]: I1001 10:45:41.043124 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" event={"ID":"8985cb77-9aeb-4c86-83b2-c93d786c4a04","Type":"ContainerDied","Data":"99289b4707c62fc27651295fbf0ad8dad630b570b04a6a3baf6fc36503a93c5a"} Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.200171 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.357625 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host\") pod \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.357855 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host" (OuterVolumeSpecName: "host") pod "8985cb77-9aeb-4c86-83b2-c93d786c4a04" (UID: "8985cb77-9aeb-4c86-83b2-c93d786c4a04"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.358119 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww87n\" (UniqueName: \"kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n\") pod \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\" (UID: \"8985cb77-9aeb-4c86-83b2-c93d786c4a04\") " Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.358787 4837 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8985cb77-9aeb-4c86-83b2-c93d786c4a04-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.363957 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n" (OuterVolumeSpecName: "kube-api-access-ww87n") pod "8985cb77-9aeb-4c86-83b2-c93d786c4a04" (UID: "8985cb77-9aeb-4c86-83b2-c93d786c4a04"). InnerVolumeSpecName "kube-api-access-ww87n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.460576 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww87n\" (UniqueName: \"kubernetes.io/projected/8985cb77-9aeb-4c86-83b2-c93d786c4a04-kube-api-access-ww87n\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:42 crc kubenswrapper[4837]: I1001 10:45:42.833937 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5m4r" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" probeResult="failure" output=< Oct 01 10:45:42 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:45:42 crc kubenswrapper[4837]: > Oct 01 10:45:43 crc kubenswrapper[4837]: I1001 10:45:43.084486 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" event={"ID":"8985cb77-9aeb-4c86-83b2-c93d786c4a04","Type":"ContainerDied","Data":"c2d3db9e62ddfad044c51a5e8d5109f8d65df45839a09acc1e7aac41cb58de57"} Oct 01 10:45:43 crc kubenswrapper[4837]: I1001 10:45:43.084880 4837 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2d3db9e62ddfad044c51a5e8d5109f8d65df45839a09acc1e7aac41cb58de57" Oct 01 10:45:43 crc kubenswrapper[4837]: I1001 10:45:43.084649 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-ct5pr" Oct 01 10:45:44 crc kubenswrapper[4837]: I1001 10:45:44.678826 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_3f958bc4-38ae-49d4-9a51-205ce949a3b8/init-config-reloader/0.log" Oct 01 10:45:45 crc kubenswrapper[4837]: I1001 10:45:45.146256 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_3f958bc4-38ae-49d4-9a51-205ce949a3b8/init-config-reloader/0.log" Oct 01 10:45:45 crc kubenswrapper[4837]: I1001 10:45:45.147239 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_3f958bc4-38ae-49d4-9a51-205ce949a3b8/alertmanager/0.log" Oct 01 10:45:45 crc kubenswrapper[4837]: I1001 10:45:45.499518 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_3f958bc4-38ae-49d4-9a51-205ce949a3b8/config-reloader/0.log" Oct 01 10:45:45 crc kubenswrapper[4837]: I1001 10:45:45.924604 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4a9fe3dd-df71-4fad-9ea3-ee024052634a/aodh-api/0.log" Oct 01 10:45:45 crc kubenswrapper[4837]: I1001 10:45:45.985953 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4a9fe3dd-df71-4fad-9ea3-ee024052634a/aodh-evaluator/0.log" Oct 01 10:45:46 crc kubenswrapper[4837]: I1001 10:45:46.170599 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4a9fe3dd-df71-4fad-9ea3-ee024052634a/aodh-listener/0.log" Oct 01 10:45:46 crc kubenswrapper[4837]: I1001 10:45:46.392515 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_4a9fe3dd-df71-4fad-9ea3-ee024052634a/aodh-notifier/0.log" Oct 01 10:45:46 crc kubenswrapper[4837]: I1001 10:45:46.666828 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54d786f8b6-vxnx7_1ece5705-7965-4476-ae5d-7ef6e80970dd/barbican-api-log/0.log" Oct 01 10:45:46 crc kubenswrapper[4837]: I1001 10:45:46.693627 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54d786f8b6-vxnx7_1ece5705-7965-4476-ae5d-7ef6e80970dd/barbican-api/0.log" Oct 01 10:45:46 crc kubenswrapper[4837]: I1001 10:45:46.965111 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69547f7db-nvq8j_deb6c8d1-7a9e-4b84-b12e-6b744c500283/barbican-keystone-listener/0.log" Oct 01 10:45:47 crc kubenswrapper[4837]: I1001 10:45:47.617681 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7bb6d5579f-w8v2b_418f6100-faf8-46ae-b22d-bf013ff3c850/barbican-worker/0.log" Oct 01 10:45:47 crc kubenswrapper[4837]: I1001 10:45:47.870187 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7bb6d5579f-w8v2b_418f6100-faf8-46ae-b22d-bf013ff3c850/barbican-worker-log/0.log" Oct 01 10:45:48 crc kubenswrapper[4837]: I1001 10:45:48.000249 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69547f7db-nvq8j_deb6c8d1-7a9e-4b84-b12e-6b744c500283/barbican-keystone-listener-log/0.log" Oct 01 10:45:48 crc kubenswrapper[4837]: I1001 10:45:48.341949 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-lhpww_18443fd2-82a7-4be3-9ad7-8be4777ec6e5/bootstrap-openstack-openstack-cell1/0.log" Oct 01 10:45:48 crc kubenswrapper[4837]: I1001 10:45:48.739219 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-networker-trbtg_7b2e57be-af14-4e5b-8751-dc110da24b25/bootstrap-openstack-openstack-networker/0.log" Oct 01 10:45:49 crc kubenswrapper[4837]: I1001 10:45:49.354275 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a41770e-165b-4d5c-9e69-8adf636ea42c/proxy-httpd/0.log" Oct 01 10:45:49 crc kubenswrapper[4837]: I1001 10:45:49.545662 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a41770e-165b-4d5c-9e69-8adf636ea42c/ceilometer-central-agent/0.log" Oct 01 10:45:49 crc kubenswrapper[4837]: I1001 10:45:49.827010 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a41770e-165b-4d5c-9e69-8adf636ea42c/sg-core/0.log" Oct 01 10:45:50 crc kubenswrapper[4837]: I1001 10:45:50.271120 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c455811f-4366-4ae6-8814-58b17ce42597/cinder-api-log/0.log" Oct 01 10:45:50 crc kubenswrapper[4837]: I1001 10:45:50.513683 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c455811f-4366-4ae6-8814-58b17ce42597/cinder-api/0.log" Oct 01 10:45:50 crc kubenswrapper[4837]: I1001 10:45:50.563996 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a41770e-165b-4d5c-9e69-8adf636ea42c/ceilometer-notification-agent/0.log" Oct 01 10:45:50 crc kubenswrapper[4837]: I1001 10:45:50.792165 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3619414f-7707-4a54-a834-bf1a1f346a14/cinder-scheduler/0.log" Oct 01 10:45:51 crc kubenswrapper[4837]: I1001 10:45:51.051913 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3619414f-7707-4a54-a834-bf1a1f346a14/probe/0.log" Oct 01 10:45:51 crc kubenswrapper[4837]: I1001 10:45:51.236757 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-hpq4d_9ccf0a84-b87b-4a54-b268-8b55d19dba49/configure-network-openstack-openstack-cell1/0.log" Oct 01 10:45:51 crc kubenswrapper[4837]: I1001 10:45:51.706423 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-networker-547hg_b189babd-cae5-47d1-bc90-c8a49033576e/configure-network-openstack-openstack-networker/0.log" Oct 01 10:45:51 crc kubenswrapper[4837]: I1001 10:45:51.774671 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-gmdpl_587efc59-4844-4bb9-8467-5c39b985fab0/configure-os-openstack-openstack-cell1/0.log" Oct 01 10:45:51 crc kubenswrapper[4837]: I1001 10:45:51.983861 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-hsszd_d5fea961-8894-40a6-a2e6-31d5b2646323/configure-os-openstack-openstack-cell1/0.log" Oct 01 10:45:52 crc kubenswrapper[4837]: I1001 10:45:52.403398 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-4qf6g_36a513c6-8a4c-4f84-b622-8e29610038ff/configure-os-openstack-openstack-networker/0.log" Oct 01 10:45:52 crc kubenswrapper[4837]: I1001 10:45:52.793797 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-snpwc_d9e207f3-0eca-47c1-aebf-de892fe662cc/configure-os-openstack-openstack-networker/0.log" Oct 01 10:45:52 crc kubenswrapper[4837]: I1001 10:45:52.836500 4837 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5m4r" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" probeResult="failure" output=< Oct 01 10:45:52 crc kubenswrapper[4837]: timeout: failed to connect service ":50051" within 1s Oct 01 10:45:52 crc kubenswrapper[4837]: > Oct 01 10:45:52 crc kubenswrapper[4837]: I1001 10:45:52.853263 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-64c4696475-qx7h5_dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546/init/0.log" Oct 01 10:45:53 crc kubenswrapper[4837]: I1001 10:45:53.078988 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:45:53 crc kubenswrapper[4837]: I1001 10:45:53.079035 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:45:53 crc kubenswrapper[4837]: I1001 10:45:53.517488 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-64c4696475-qx7h5_dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546/init/0.log" Oct 01 10:45:54 crc kubenswrapper[4837]: I1001 10:45:54.012183 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-nbbwp_4a18e58a-b31f-43b9-8fac-5d55c9e65621/download-cache-openstack-openstack-cell1/0.log" Oct 01 10:45:54 crc kubenswrapper[4837]: I1001 10:45:54.352120 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-networker-wqxpl_f90a4c39-8e54-4d88-8e72-b8ad8caadd2f/download-cache-openstack-openstack-networker/0.log" Oct 01 10:45:54 crc kubenswrapper[4837]: I1001 10:45:54.904303 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b3b9c094-9ba3-4e91-9ec3-168180bc56f0/glance-httpd/0.log" Oct 01 10:45:54 crc kubenswrapper[4837]: I1001 10:45:54.999180 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b3b9c094-9ba3-4e91-9ec3-168180bc56f0/glance-log/0.log" Oct 01 10:45:55 crc kubenswrapper[4837]: I1001 10:45:55.350977 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-ct5pr"] Oct 01 10:45:55 crc kubenswrapper[4837]: I1001 10:45:55.368385 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-ct5pr"] Oct 01 10:45:55 crc kubenswrapper[4837]: I1001 10:45:55.387990 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_50f83292-0b00-4281-8ed3-e072a08a6c5d/glance-httpd/0.log" Oct 01 10:45:55 crc kubenswrapper[4837]: I1001 10:45:55.428403 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_50f83292-0b00-4281-8ed3-e072a08a6c5d/glance-log/0.log" Oct 01 10:45:55 crc kubenswrapper[4837]: I1001 10:45:55.829349 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8985cb77-9aeb-4c86-83b2-c93d786c4a04" path="/var/lib/kubelet/pods/8985cb77-9aeb-4c86-83b2-c93d786c4a04/volumes" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.456661 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-b8cc69c8f-zc64n_d9393af6-558d-4e75-8602-f9e8d09ceb7f/heat-engine/0.log" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.622509 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-q882j"] Oct 01 10:45:56 crc kubenswrapper[4837]: E1001 10:45:56.623558 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8985cb77-9aeb-4c86-83b2-c93d786c4a04" containerName="container-00" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.623579 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="8985cb77-9aeb-4c86-83b2-c93d786c4a04" containerName="container-00" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.623858 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="8985cb77-9aeb-4c86-83b2-c93d786c4a04" containerName="container-00" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.625218 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.670467 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt2x5\" (UniqueName: \"kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.670531 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.772482 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt2x5\" (UniqueName: \"kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.772531 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.772797 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.806911 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt2x5\" (UniqueName: \"kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5\") pod \"crc-debug-q882j\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.947243 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:56 crc kubenswrapper[4837]: I1001 10:45:56.950748 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5dc5f6c468-wpmkd_da4dea0d-4b10-456a-9704-a6012b34bd40/heat-api/0.log" Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.231483 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-575bccb888-ndbhd_df71e942-12b3-4d00-8238-bf93a99ab3a9/horizon/0.log" Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.298713 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-q882j" event={"ID":"53345ea0-5f2c-495b-be60-10198ea0cac5","Type":"ContainerStarted","Data":"45203494b39886a098a179d36d4f25688eec045680913ddff0cb068609763800"} Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.299099 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-q882j" event={"ID":"53345ea0-5f2c-495b-be60-10198ea0cac5","Type":"ContainerStarted","Data":"f9246499c7f09747369e1692e0629eabae9d18eae8a3804437645c2555a04be3"} Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.328999 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-77ffb6c584-qdkpl_6c720536-3f20-4c58-b914-ddbd3c52645b/heat-cfnapi/0.log" Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.331720 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ntfc2/crc-debug-q882j" podStartSLOduration=1.331699699 podStartE2EDuration="1.331699699s" podCreationTimestamp="2025-10-01 10:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:45:57.31390417 +0000 UTC m=+13214.155511625" watchObservedRunningTime="2025-10-01 10:45:57.331699699 +0000 UTC m=+13214.173307154" Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.567983 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-zvqxc_03068cc4-7997-4279-98d1-1e6705b0bd75/install-certs-openstack-openstack-cell1/0.log" Oct 01 10:45:57 crc kubenswrapper[4837]: I1001 10:45:57.876978 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-networker-tvk69_5d158e44-e94a-46e6-8fd5-82e31f63aff0/install-certs-openstack-openstack-networker/0.log" Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.163868 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-5bp2d_5040522c-817c-455f-a58c-5becbf01ac90/install-os-openstack-openstack-cell1/0.log" Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.289539 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-575bccb888-ndbhd_df71e942-12b3-4d00-8238-bf93a99ab3a9/horizon-log/0.log" Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.317592 4837 generic.go:334] "Generic (PLEG): container finished" podID="53345ea0-5f2c-495b-be60-10198ea0cac5" containerID="45203494b39886a098a179d36d4f25688eec045680913ddff0cb068609763800" exitCode=0 Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.317635 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/crc-debug-q882j" event={"ID":"53345ea0-5f2c-495b-be60-10198ea0cac5","Type":"ContainerDied","Data":"45203494b39886a098a179d36d4f25688eec045680913ddff0cb068609763800"} Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.459516 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-networker-gsgxm_89f2b30f-c5e3-4a62-9092-4b00ea001f3c/install-os-openstack-openstack-networker/0.log" Oct 01 10:45:58 crc kubenswrapper[4837]: I1001 10:45:58.544554 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-64c4696475-qx7h5_dc42f6d7-dda3-4968-a4fc-d9b5ec2d5546/dnsmasq-dns/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.142784 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29321821-xqq67_bc041921-e6b3-4855-9685-8855f4efe2db/keystone-cron/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.224985 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29321881-q6spk_bce0fa96-7165-40ce-b234-74e76950f708/keystone-cron/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.400494 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-9788869bb-qfvkm_d4d8d5b0-1282-4dff-92ff-5a79de21571c/keystone-api/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.445916 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.482720 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-q882j"] Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.497007 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ntfc2/crc-debug-q882j"] Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.529768 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt2x5\" (UniqueName: \"kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5\") pod \"53345ea0-5f2c-495b-be60-10198ea0cac5\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.529868 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host\") pod \"53345ea0-5f2c-495b-be60-10198ea0cac5\" (UID: \"53345ea0-5f2c-495b-be60-10198ea0cac5\") " Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.530000 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host" (OuterVolumeSpecName: "host") pod "53345ea0-5f2c-495b-be60-10198ea0cac5" (UID: "53345ea0-5f2c-495b-be60-10198ea0cac5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.530451 4837 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/53345ea0-5f2c-495b-be60-10198ea0cac5-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.533223 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_ecafde02-447b-483c-bbe5-218cf0d8f772/kube-state-metrics/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.537805 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5" (OuterVolumeSpecName: "kube-api-access-tt2x5") pod "53345ea0-5f2c-495b-be60-10198ea0cac5" (UID: "53345ea0-5f2c-495b-be60-10198ea0cac5"). InnerVolumeSpecName "kube-api-access-tt2x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.633249 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt2x5\" (UniqueName: \"kubernetes.io/projected/53345ea0-5f2c-495b-be60-10198ea0cac5-kube-api-access-tt2x5\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.722456 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-7lfc2_442bd19f-fbab-4863-a8de-098783ec10e9/libvirt-openstack-openstack-cell1/0.log" Oct 01 10:45:59 crc kubenswrapper[4837]: I1001 10:45:59.827114 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53345ea0-5f2c-495b-be60-10198ea0cac5" path="/var/lib/kubelet/pods/53345ea0-5f2c-495b-be60-10198ea0cac5/volumes" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.162947 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd98cb97c-4l4pg_c629d2d6-9a91-4a51-9504-1524ee0bc8ed/neutron-httpd/0.log" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.255838 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-2c4mx_41ead075-9a6e-4f49-812a-9c8cc8937c22/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.337631 4837 scope.go:117] "RemoveContainer" containerID="45203494b39886a098a179d36d4f25688eec045680913ddff0cb068609763800" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.337795 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/crc-debug-q882j" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.504815 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-xdctt_24d911bc-6b98-4fbf-a326-89be185e859e/neutron-metadata-openstack-openstack-cell1/0.log" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.578119 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd98cb97c-4l4pg_c629d2d6-9a91-4a51-9504-1524ee0bc8ed/neutron-api/0.log" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.836360 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-networker-thkv6_ffe682e2-2611-4971-a163-a2adfd4570d7/neutron-metadata-openstack-openstack-networker/0.log" Oct 01 10:46:00 crc kubenswrapper[4837]: I1001 10:46:00.965581 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-zgx57_a8855094-af9a-42ba-9931-55420616b36d/neutron-sriov-openstack-openstack-cell1/0.log" Oct 01 10:46:01 crc kubenswrapper[4837]: I1001 10:46:01.447084 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7a7c8913-69d5-47a0-82c1-6136fc655ba6/nova-cell0-conductor-conductor/0.log" Oct 01 10:46:01 crc kubenswrapper[4837]: I1001 10:46:01.600819 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_40228c63-32a1-4ea4-9e2f-0849056889e5/nova-api-log/0.log" Oct 01 10:46:01 crc kubenswrapper[4837]: I1001 10:46:01.760483 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a532e53d-2421-4f9a-b16e-1bbe87b9a7d7/nova-cell1-conductor-conductor/0.log" Oct 01 10:46:01 crc kubenswrapper[4837]: I1001 10:46:01.844347 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:46:01 crc kubenswrapper[4837]: I1001 10:46:01.905780 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.017189 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ba4d4a69-79d0-4f8b-83db-319fd0aca696/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.092723 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.150073 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_40228c63-32a1-4ea4-9e2f-0849056889e5/nova-api-api/0.log" Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.297539 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellxvxw5_956da144-c603-4569-b09e-0f4bce2d6c15/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.495163 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-rjp2f_58f455af-4ff9-4beb-a59f-8ccfc4860c67/nova-cell1-openstack-openstack-cell1/0.log" Oct 01 10:46:02 crc kubenswrapper[4837]: I1001 10:46:02.695084 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_db57ec59-8011-46c9-b1ba-e113bb6b8137/nova-metadata-log/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.065596 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_001d3974-ad2f-4ef1-880c-ce4e32600f24/mysql-bootstrap/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.133245 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7fcdf164-8ca8-44f8-b44b-763300203256/nova-scheduler-scheduler/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.371013 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x5m4r" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" containerID="cri-o://2439747f810c4a0fc9a136337b338d25cf27821b26492a5ef34a1ede74f44010" gracePeriod=2 Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.403606 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_001d3974-ad2f-4ef1-880c-ce4e32600f24/mysql-bootstrap/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.414640 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_001d3974-ad2f-4ef1-880c-ce4e32600f24/galera/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.703532 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5bced997-a865-4e08-b41d-53f9ed5ea466/mysql-bootstrap/0.log" Oct 01 10:46:03 crc kubenswrapper[4837]: I1001 10:46:03.964731 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5bced997-a865-4e08-b41d-53f9ed5ea466/mysql-bootstrap/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.037360 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5bced997-a865-4e08-b41d-53f9ed5ea466/galera/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.289677 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_de8d83c2-5603-4289-bff7-5121a228c96b/openstack-network-exporter/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.310130 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_87d45b64-bd5f-4828-900a-0077a3f6b59f/openstackclient/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.319228 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_db57ec59-8011-46c9-b1ba-e113bb6b8137/nova-metadata-metadata/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.386897 4837 generic.go:334] "Generic (PLEG): container finished" podID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerID="2439747f810c4a0fc9a136337b338d25cf27821b26492a5ef34a1ede74f44010" exitCode=0 Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.386939 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerDied","Data":"2439747f810c4a0fc9a136337b338d25cf27821b26492a5ef34a1ede74f44010"} Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.643396 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_de8d83c2-5603-4289-bff7-5121a228c96b/ovn-northd/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.673973 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-2d8vv_907ebdee-b8d2-47ef-b934-615ec125f46b/ovn-openstack-openstack-cell1/0.log" Oct 01 10:46:04 crc kubenswrapper[4837]: I1001 10:46:04.995081 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_fde602a4-cc23-4e59-bb85-c3923a628294/openstack-network-exporter/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.014067 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.056284 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content\") pod \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.056400 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities\") pod \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.059643 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vbcs\" (UniqueName: \"kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs\") pod \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\" (UID: \"b7fd680a-f3d7-4292-b50c-cdd8938557c6\") " Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.061145 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities" (OuterVolumeSpecName: "utilities") pod "b7fd680a-f3d7-4292-b50c-cdd8938557c6" (UID: "b7fd680a-f3d7-4292-b50c-cdd8938557c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.086886 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs" (OuterVolumeSpecName: "kube-api-access-2vbcs") pod "b7fd680a-f3d7-4292-b50c-cdd8938557c6" (UID: "b7fd680a-f3d7-4292-b50c-cdd8938557c6"). InnerVolumeSpecName "kube-api-access-2vbcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.142009 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_fde602a4-cc23-4e59-bb85-c3923a628294/ovsdbserver-nb/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.166487 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.166522 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vbcs\" (UniqueName: \"kubernetes.io/projected/b7fd680a-f3d7-4292-b50c-cdd8938557c6-kube-api-access-2vbcs\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.201381 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7fd680a-f3d7-4292-b50c-cdd8938557c6" (UID: "b7fd680a-f3d7-4292-b50c-cdd8938557c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.240007 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-networker-nvhcc_820b0167-6b33-4527-a211-e1ccbcf19a65/ovn-openstack-openstack-networker/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.268488 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7fd680a-f3d7-4292-b50c-cdd8938557c6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.358860 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_75755bf0-420a-42fd-8d67-0df40028d9ed/openstack-network-exporter/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.406390 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5m4r" event={"ID":"b7fd680a-f3d7-4292-b50c-cdd8938557c6","Type":"ContainerDied","Data":"b75a433bc0b47c3232d05001538bfcb7932ccc4c089c401f27eac4da0ca98e26"} Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.406442 4837 scope.go:117] "RemoveContainer" containerID="2439747f810c4a0fc9a136337b338d25cf27821b26492a5ef34a1ede74f44010" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.406595 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5m4r" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.441265 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.451885 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_75755bf0-420a-42fd-8d67-0df40028d9ed/ovsdbserver-nb/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.454053 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x5m4r"] Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.468093 4837 scope.go:117] "RemoveContainer" containerID="133655a0504063b9c995de189ff9a4e3e3d903e3971d68f3309c04c7e24fa749" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.506859 4837 scope.go:117] "RemoveContainer" containerID="9c3ad630b4d5b753f457fd115c6a3e069cbfc7c07f09643c0314046253e634b1" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.619677 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b/openstack-network-exporter/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.699813 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_ca1a22b9-7313-4f9f-a9f4-b2efe0aa508b/ovsdbserver-nb/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.834975 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" path="/var/lib/kubelet/pods/b7fd680a-f3d7-4292-b50c-cdd8938557c6/volumes" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.891434 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416/ovsdbserver-sb/0.log" Oct 01 10:46:05 crc kubenswrapper[4837]: I1001 10:46:05.938767 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1fbaf3d0-9cb4-4e03-bd2e-1ab9ea38d416/openstack-network-exporter/0.log" Oct 01 10:46:06 crc kubenswrapper[4837]: I1001 10:46:06.374136 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_679a277e-9e24-4a5a-ad48-6ba934d0a55a/ovsdbserver-sb/0.log" Oct 01 10:46:06 crc kubenswrapper[4837]: I1001 10:46:06.412084 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_679a277e-9e24-4a5a-ad48-6ba934d0a55a/openstack-network-exporter/0.log" Oct 01 10:46:06 crc kubenswrapper[4837]: I1001 10:46:06.517966 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_680fbe7d-0b4f-4755-b18e-21436e4c0530/openstack-network-exporter/0.log" Oct 01 10:46:06 crc kubenswrapper[4837]: I1001 10:46:06.783950 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_680fbe7d-0b4f-4755-b18e-21436e4c0530/ovsdbserver-sb/0.log" Oct 01 10:46:06 crc kubenswrapper[4837]: I1001 10:46:06.996316 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85d78dc59b-6dx5d_302856a7-e2da-4d70-b4f9-21b24570eeaa/placement-api/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.130455 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cdqjvh_d99bc7f6-69da-47d6-8a2b-0efeed9445cf/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.249318 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85d78dc59b-6dx5d_302856a7-e2da-4d70-b4f9-21b24570eeaa/placement-log/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.364718 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-n2924t_e9f6cd02-35a4-4528-abd6-70d505abe8c3/pre-adoption-validation-openstack-pre-adoption-openstack-networ/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.587913 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dfb858f6-b1bb-41ee-a7be-ecef06894512/init-config-reloader/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.896312 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dfb858f6-b1bb-41ee-a7be-ecef06894512/prometheus/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.901728 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dfb858f6-b1bb-41ee-a7be-ecef06894512/thanos-sidecar/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.923891 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dfb858f6-b1bb-41ee-a7be-ecef06894512/init-config-reloader/0.log" Oct 01 10:46:07 crc kubenswrapper[4837]: I1001 10:46:07.926841 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_dfb858f6-b1bb-41ee-a7be-ecef06894512/config-reloader/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.219121 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cf357d6-04ea-4f07-a430-32eaba72e2da/setup-container/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.629926 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cda9191-47f1-4a0e-9c7a-32775705c5a5/setup-container/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.643638 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cf357d6-04ea-4f07-a430-32eaba72e2da/setup-container/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.746880 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cf357d6-04ea-4f07-a430-32eaba72e2da/rabbitmq/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.932537 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cda9191-47f1-4a0e-9c7a-32775705c5a5/setup-container/0.log" Oct 01 10:46:08 crc kubenswrapper[4837]: I1001 10:46:08.972725 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3cda9191-47f1-4a0e-9c7a-32775705c5a5/rabbitmq/0.log" Oct 01 10:46:09 crc kubenswrapper[4837]: I1001 10:46:09.138537 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-n94sg_b2540bf2-f9ab-41c6-a5ca-bee9e886cabd/reboot-os-openstack-openstack-cell1/0.log" Oct 01 10:46:09 crc kubenswrapper[4837]: I1001 10:46:09.338566 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-networker-dz5sw_c6106757-a810-46cd-8b2c-9e455dc9f95b/reboot-os-openstack-openstack-networker/0.log" Oct 01 10:46:09 crc kubenswrapper[4837]: I1001 10:46:09.451682 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-swmd5_08a6e6cb-9507-4d0e-b23a-5816180342a8/run-os-openstack-openstack-cell1/0.log" Oct 01 10:46:09 crc kubenswrapper[4837]: I1001 10:46:09.953651 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-networker-jffww_5f6065c8-1d4b-4302-84e4-7e30860f1864/run-os-openstack-openstack-networker/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.034440 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-tlfzz_455fb732-e3bd-47a1-9428-b6a903332b76/ssh-known-hosts-openstack/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.323644 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-45l5p_5fc49e27-d900-4566-a796-c1d6e72e2444/swift-ring-rebalance/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.449262 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b44b6c8b7-fh9pg_2b4b56f8-7b62-4209-9683-8a6c842eabe5/proxy-server/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.567154 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b44b6c8b7-fh9pg_2b4b56f8-7b62-4209-9683-8a6c842eabe5/proxy-httpd/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.670297 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-debug-lkpbl_5d2b6854-fb7d-4a91-ace2-1c25c131f9fb/swift-ring-rebalance/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.862996 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/account-reaper/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.903252 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/account-auditor/0.log" Oct 01 10:46:10 crc kubenswrapper[4837]: I1001 10:46:10.977093 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/account-replicator/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.141918 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/container-auditor/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.151817 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/container-replicator/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.191123 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/account-server/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.359830 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/container-updater/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.427523 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/object-auditor/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.547296 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/object-expirer/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.704833 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/object-replicator/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.826495 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/object-updater/0.log" Oct 01 10:46:11 crc kubenswrapper[4837]: I1001 10:46:11.940311 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/container-server/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.052032 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/swift-recon-cron/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.289741 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/object-server/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.356314 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/account-auditor/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.438398 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/account-reaper/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.481293 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f5622e0a-790d-453a-bdcc-86e778a90b61/rsync/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.577067 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/account-replicator/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.784031 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/container-auditor/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.802489 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/container-replicator/0.log" Oct 01 10:46:12 crc kubenswrapper[4837]: I1001 10:46:12.876684 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/account-server/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.007577 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/container-updater/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.126566 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/object-auditor/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.126972 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/object-expirer/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.318496 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/object-replicator/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.411605 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/object-updater/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.482602 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/container-server/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.638816 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/swift-recon-cron/0.log" Oct 01 10:46:13 crc kubenswrapper[4837]: I1001 10:46:13.954251 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/account-auditor/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.047578 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/object-server/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.101252 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/account-reaper/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.200002 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_7b4549e0-58e4-4bf6-af2b-6fbad629511c/rsync/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.252503 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/account-replicator/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.339075 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/container-auditor/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.583296 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/container-replicator/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.633178 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/account-server/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.754222 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/container-updater/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.862442 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/object-auditor/0.log" Oct 01 10:46:14 crc kubenswrapper[4837]: I1001 10:46:14.933106 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/object-expirer/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.075544 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/object-replicator/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.094061 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/container-server/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.202927 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/object-updater/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.395488 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/swift-recon-cron/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.754209 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-dg7g9_177a9873-846f-4c66-8dda-abe9d4874b96/telemetry-openstack-openstack-cell1/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.766192 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/object-server/0.log" Oct 01 10:46:15 crc kubenswrapper[4837]: I1001 10:46:15.931765 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_ac458f22-6672-45bf-89f6-d19c35debb4f/rsync/0.log" Oct 01 10:46:16 crc kubenswrapper[4837]: I1001 10:46:16.178088 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c57e86bb-3dca-4c86-938a-4072fa900d0c/test-operator-logs-container/0.log" Oct 01 10:46:16 crc kubenswrapper[4837]: I1001 10:46:16.279647 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_62fa4c75-1944-4f6f-8efe-8463b065c5c5/tempest-tests-tempest-tests-runner/0.log" Oct 01 10:46:16 crc kubenswrapper[4837]: I1001 10:46:16.575737 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-rf89t_3e10153b-fcd7-432f-95c2-c240f6b53944/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 01 10:46:16 crc kubenswrapper[4837]: I1001 10:46:16.643873 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-networker-p5brm_5fc52d12-9bf2-4649-9ea5-68d196c129a4/tripleo-cleanup-tripleo-cleanup-openstack-networker/0.log" Oct 01 10:46:16 crc kubenswrapper[4837]: I1001 10:46:16.843304 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-7nqgj_85317dc9-c42f-421a-9101-787a19352703/validate-network-openstack-openstack-cell1/0.log" Oct 01 10:46:17 crc kubenswrapper[4837]: I1001 10:46:17.086237 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-networker-mfk6z_e4fdd3dd-4022-4ec7-bc01-d489e1127fb0/validate-network-openstack-openstack-networker/0.log" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.079231 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.079595 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.079653 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.080611 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.080706 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" gracePeriod=600 Oct 01 10:46:23 crc kubenswrapper[4837]: E1001 10:46:23.218345 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.662021 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" exitCode=0 Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.662062 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa"} Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.662091 4837 scope.go:117] "RemoveContainer" containerID="5e9c7d53768d0e5bca07cf23674d4e1c29c1f8c3d51297aa483c9be83793c304" Oct 01 10:46:23 crc kubenswrapper[4837]: I1001 10:46:23.662731 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:46:23 crc kubenswrapper[4837]: E1001 10:46:23.662966 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:46:28 crc kubenswrapper[4837]: I1001 10:46:28.501685 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e98b8d09-2148-42d2-aeef-650cc3b41feb/memcached/0.log" Oct 01 10:46:35 crc kubenswrapper[4837]: I1001 10:46:35.815794 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:46:35 crc kubenswrapper[4837]: E1001 10:46:35.816429 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.154916 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:42 crc kubenswrapper[4837]: E1001 10:46:42.155832 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="extract-content" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.155844 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="extract-content" Oct 01 10:46:42 crc kubenswrapper[4837]: E1001 10:46:42.155858 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53345ea0-5f2c-495b-be60-10198ea0cac5" containerName="container-00" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.155864 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="53345ea0-5f2c-495b-be60-10198ea0cac5" containerName="container-00" Oct 01 10:46:42 crc kubenswrapper[4837]: E1001 10:46:42.155882 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.155888 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" Oct 01 10:46:42 crc kubenswrapper[4837]: E1001 10:46:42.155918 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="extract-utilities" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.155926 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="extract-utilities" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.156123 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fd680a-f3d7-4292-b50c-cdd8938557c6" containerName="registry-server" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.156153 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="53345ea0-5f2c-495b-be60-10198ea0cac5" containerName="container-00" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.161873 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.211686 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.317189 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.317338 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8d9q\" (UniqueName: \"kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.317392 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.419279 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.419347 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8d9q\" (UniqueName: \"kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.419367 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.419804 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.419942 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.443018 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8d9q\" (UniqueName: \"kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q\") pod \"redhat-marketplace-v2dlh\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:42 crc kubenswrapper[4837]: I1001 10:46:42.484448 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:43 crc kubenswrapper[4837]: I1001 10:46:43.241604 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:43 crc kubenswrapper[4837]: I1001 10:46:43.840423 4837 generic.go:334] "Generic (PLEG): container finished" podID="9156916c-49ea-47af-88e4-aa7941667362" containerID="0250db8a3bab7ce799464fcd57b2495b643fbcf1b3573b3adda4bf40412adfaa" exitCode=0 Oct 01 10:46:43 crc kubenswrapper[4837]: I1001 10:46:43.841194 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerDied","Data":"0250db8a3bab7ce799464fcd57b2495b643fbcf1b3573b3adda4bf40412adfaa"} Oct 01 10:46:43 crc kubenswrapper[4837]: I1001 10:46:43.842356 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerStarted","Data":"d721e01cc0df548385515ab132d91ff21d7b370fd595c13f848171c6426560b7"} Oct 01 10:46:44 crc kubenswrapper[4837]: I1001 10:46:44.856029 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerStarted","Data":"7653f1c02fb4916e5667e867e85aa4b2fc990a2405daca8cb00a727f74cea571"} Oct 01 10:46:45 crc kubenswrapper[4837]: I1001 10:46:45.868332 4837 generic.go:334] "Generic (PLEG): container finished" podID="9156916c-49ea-47af-88e4-aa7941667362" containerID="7653f1c02fb4916e5667e867e85aa4b2fc990a2405daca8cb00a727f74cea571" exitCode=0 Oct 01 10:46:45 crc kubenswrapper[4837]: I1001 10:46:45.868409 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerDied","Data":"7653f1c02fb4916e5667e867e85aa4b2fc990a2405daca8cb00a727f74cea571"} Oct 01 10:46:46 crc kubenswrapper[4837]: I1001 10:46:46.881217 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerStarted","Data":"bce23e5bf12e0b2e54df8cc9badfd0b89ffad5043c145487ec860d2a53a28d4f"} Oct 01 10:46:46 crc kubenswrapper[4837]: I1001 10:46:46.913020 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v2dlh" podStartSLOduration=2.405157476 podStartE2EDuration="4.913004521s" podCreationTimestamp="2025-10-01 10:46:42 +0000 UTC" firstStartedPulling="2025-10-01 10:46:43.84383484 +0000 UTC m=+13260.685442295" lastFinishedPulling="2025-10-01 10:46:46.351681855 +0000 UTC m=+13263.193289340" observedRunningTime="2025-10-01 10:46:46.905064145 +0000 UTC m=+13263.746671600" watchObservedRunningTime="2025-10-01 10:46:46.913004521 +0000 UTC m=+13263.754611976" Oct 01 10:46:48 crc kubenswrapper[4837]: I1001 10:46:48.817760 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:46:48 crc kubenswrapper[4837]: E1001 10:46:48.818164 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:46:50 crc kubenswrapper[4837]: I1001 10:46:50.777976 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-rqr6f_ebdcaccc-170c-40ff-9b06-57a1188cf20b/kube-rbac-proxy/0.log" Oct 01 10:46:50 crc kubenswrapper[4837]: I1001 10:46:50.963004 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-rqr6f_ebdcaccc-170c-40ff-9b06-57a1188cf20b/manager/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.075170 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-lm9xd_06ebbbd0-256b-4637-b6e1-ba5adab9bd40/kube-rbac-proxy/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.175824 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-lm9xd_06ebbbd0-256b-4637-b6e1-ba5adab9bd40/manager/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.226097 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-lq7sr_e5e81392-cf50-43ab-b86a-e8881d5043f7/kube-rbac-proxy/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.323232 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-lq7sr_e5e81392-cf50-43ab-b86a-e8881d5043f7/manager/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.420383 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/util/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.664897 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/pull/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.674063 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/util/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.713919 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/pull/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.936959 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/extract/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.961128 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/util/0.log" Oct 01 10:46:51 crc kubenswrapper[4837]: I1001 10:46:51.983225 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_edfb8f1e3757614357120847610749d1bf40332c7e495947e8dcac02c1bmbgg_deffc359-4840-433e-999e-fa7787e97ed2/pull/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.374796 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-wnqvm_750df3b7-ef90-4f83-99f9-ebb54a6a159b/kube-rbac-proxy/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.484562 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.484613 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.526417 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-wnqvm_750df3b7-ef90-4f83-99f9-ebb54a6a159b/manager/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.555617 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.618834 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-95dns_426cfad6-045e-48e1-82e4-ce9b9891dd0a/kube-rbac-proxy/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.653451 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-95dns_426cfad6-045e-48e1-82e4-ce9b9891dd0a/manager/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.850456 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-rz6jg_0cd7b250-33bc-45fc-be0d-e4d6d28e450f/manager/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.865978 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-rz6jg_0cd7b250-33bc-45fc-be0d-e4d6d28e450f/kube-rbac-proxy/0.log" Oct 01 10:46:52 crc kubenswrapper[4837]: I1001 10:46:52.993353 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.009664 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-tlhpd_5d9a421b-7104-462a-8a01-8bc62522a543/kube-rbac-proxy/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.221662 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f589bc7f7-wf96x_f987881b-66b4-49fe-a171-15813e9225f5/kube-rbac-proxy/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.261179 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f589bc7f7-wf96x_f987881b-66b4-49fe-a171-15813e9225f5/manager/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.359534 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-tlhpd_5d9a421b-7104-462a-8a01-8bc62522a543/manager/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.495420 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-6hgks_b537c4f7-697f-4856-b7f9-98a55150cc9f/kube-rbac-proxy/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.539680 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.608786 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-6hgks_b537c4f7-697f-4856-b7f9-98a55150cc9f/manager/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.741362 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-l4pdl_3efcdf55-46eb-446e-933a-e086f3a67aab/kube-rbac-proxy/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.790518 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-l4pdl_3efcdf55-46eb-446e-933a-e086f3a67aab/manager/0.log" Oct 01 10:46:53 crc kubenswrapper[4837]: I1001 10:46:53.863257 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-mgm24_e6fb38c5-e91c-4229-9aa5-93729ffab045/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.012990 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-mgm24_e6fb38c5-e91c-4229-9aa5-93729ffab045/manager/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.053418 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6b96467f46-kgf8r_fef6cdf1-8dbe-47e9-8d8a-143710d954ec/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.198170 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6b96467f46-kgf8r_fef6cdf1-8dbe-47e9-8d8a-143710d954ec/manager/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.316491 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79f9fc9fd8-p978w_76e4d2ee-d38f-433c-945c-29f36594a338/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.488199 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79f9fc9fd8-p978w_76e4d2ee-d38f-433c-945c-29f36594a338/manager/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.550439 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fb7d6b8bf-2xtdw_2ed39459-5a70-4c13-aacd-966aae697780/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.646539 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fb7d6b8bf-2xtdw_2ed39459-5a70-4c13-aacd-966aae697780/manager/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.769907 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q_648417c5-8d92-4426-81db-052bcb20cbc6/manager/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.805864 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-cc9cf7895-jjc4q_648417c5-8d92-4426-81db-052bcb20cbc6/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.840021 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8648d97544-8vjp6_7de7f612-a5bd-4f2c-a146-d5f49ba757af/kube-rbac-proxy/0.log" Oct 01 10:46:54 crc kubenswrapper[4837]: I1001 10:46:54.955567 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v2dlh" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="registry-server" containerID="cri-o://bce23e5bf12e0b2e54df8cc9badfd0b89ffad5043c145487ec860d2a53a28d4f" gracePeriod=2 Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.019448 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8d79f897-jm6br_f53920b8-af09-4740-a121-b4b2207bba70/kube-rbac-proxy/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.255195 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8d79f897-jm6br_f53920b8-af09-4740-a121-b4b2207bba70/operator/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.523319 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-6bm6v_969f8b13-b131-4fad-84ff-6b092fcc43e9/kube-rbac-proxy/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.670636 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wvb74_0dcc771f-a525-479f-b679-0783f280bb7a/registry-server/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.750590 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-6bm6v_969f8b13-b131-4fad-84ff-6b092fcc43e9/manager/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.859242 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-822wp_d4fc2a84-086a-4644-b354-2ef743bf1e7f/kube-rbac-proxy/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.960762 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-822wp_d4fc2a84-086a-4644-b354-2ef743bf1e7f/manager/0.log" Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.967434 4837 generic.go:334] "Generic (PLEG): container finished" podID="9156916c-49ea-47af-88e4-aa7941667362" containerID="bce23e5bf12e0b2e54df8cc9badfd0b89ffad5043c145487ec860d2a53a28d4f" exitCode=0 Oct 01 10:46:55 crc kubenswrapper[4837]: I1001 10:46:55.967475 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerDied","Data":"bce23e5bf12e0b2e54df8cc9badfd0b89ffad5043c145487ec860d2a53a28d4f"} Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.072912 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-s8xw2_421ed6b2-06bf-421e-a871-ec953b32bc93/operator/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.185617 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-657c6b68c7-fddfz_70c21bc2-f519-4551-bf8d-669df6064834/kube-rbac-proxy/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.353956 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-657c6b68c7-fddfz_70c21bc2-f519-4551-bf8d-669df6064834/manager/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.463995 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.562768 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmqwh_2297d6d0-957e-4d04-9194-d9d7c120c3f0/kube-rbac-proxy/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.609444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content\") pod \"9156916c-49ea-47af-88e4-aa7941667362\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.609503 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8d9q\" (UniqueName: \"kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q\") pod \"9156916c-49ea-47af-88e4-aa7941667362\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.609583 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities\") pod \"9156916c-49ea-47af-88e4-aa7941667362\" (UID: \"9156916c-49ea-47af-88e4-aa7941667362\") " Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.613502 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities" (OuterVolumeSpecName: "utilities") pod "9156916c-49ea-47af-88e4-aa7941667362" (UID: "9156916c-49ea-47af-88e4-aa7941667362"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.620117 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q" (OuterVolumeSpecName: "kube-api-access-r8d9q") pod "9156916c-49ea-47af-88e4-aa7941667362" (UID: "9156916c-49ea-47af-88e4-aa7941667362"). InnerVolumeSpecName "kube-api-access-r8d9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.627834 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9156916c-49ea-47af-88e4-aa7941667362" (UID: "9156916c-49ea-47af-88e4-aa7941667362"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.682539 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb97fcf96-vcm95_0a3d4dda-cb7c-48ae-95ee-f6eebe98745e/kube-rbac-proxy/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.711517 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.711551 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8d9q\" (UniqueName: \"kubernetes.io/projected/9156916c-49ea-47af-88e4-aa7941667362-kube-api-access-r8d9q\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.711562 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9156916c-49ea-47af-88e4-aa7941667362-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.803808 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb97fcf96-vcm95_0a3d4dda-cb7c-48ae-95ee-f6eebe98745e/manager/0.log" Oct 01 10:46:56 crc kubenswrapper[4837]: I1001 10:46:56.916425 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75756dd4d9-vx2m2_414c531b-a0c8-4cb4-a997-77706c2f8e7c/kube-rbac-proxy/0.log" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.005081 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v2dlh" event={"ID":"9156916c-49ea-47af-88e4-aa7941667362","Type":"ContainerDied","Data":"d721e01cc0df548385515ab132d91ff21d7b370fd595c13f848171c6426560b7"} Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.005174 4837 scope.go:117] "RemoveContainer" containerID="bce23e5bf12e0b2e54df8cc9badfd0b89ffad5043c145487ec860d2a53a28d4f" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.005452 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v2dlh" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.035267 4837 scope.go:117] "RemoveContainer" containerID="7653f1c02fb4916e5667e867e85aa4b2fc990a2405daca8cb00a727f74cea571" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.076803 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.080294 4837 scope.go:117] "RemoveContainer" containerID="0250db8a3bab7ce799464fcd57b2495b643fbcf1b3573b3adda4bf40412adfaa" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.121084 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v2dlh"] Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.143984 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75756dd4d9-vx2m2_414c531b-a0c8-4cb4-a997-77706c2f8e7c/manager/0.log" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.201170 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-cmqwh_2297d6d0-957e-4d04-9194-d9d7c120c3f0/manager/0.log" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.837279 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9156916c-49ea-47af-88e4-aa7941667362" path="/var/lib/kubelet/pods/9156916c-49ea-47af-88e4-aa7941667362/volumes" Oct 01 10:46:57 crc kubenswrapper[4837]: I1001 10:46:57.949647 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8648d97544-8vjp6_7de7f612-a5bd-4f2c-a146-d5f49ba757af/manager/0.log" Oct 01 10:46:59 crc kubenswrapper[4837]: I1001 10:46:59.816304 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:46:59 crc kubenswrapper[4837]: E1001 10:46:59.817019 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:47:13 crc kubenswrapper[4837]: I1001 10:47:13.354303 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6vht5_7b8012dc-8975-4a9d-86c3-436e7fca87fa/control-plane-machine-set-operator/0.log" Oct 01 10:47:13 crc kubenswrapper[4837]: I1001 10:47:13.598912 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-822f9_d5baef47-d279-4a19-88f6-271f843d6594/machine-api-operator/0.log" Oct 01 10:47:13 crc kubenswrapper[4837]: I1001 10:47:13.608858 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-822f9_d5baef47-d279-4a19-88f6-271f843d6594/kube-rbac-proxy/0.log" Oct 01 10:47:14 crc kubenswrapper[4837]: I1001 10:47:14.816729 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:47:14 crc kubenswrapper[4837]: E1001 10:47:14.817249 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:47:26 crc kubenswrapper[4837]: I1001 10:47:26.816843 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:47:26 crc kubenswrapper[4837]: E1001 10:47:26.818457 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:47:29 crc kubenswrapper[4837]: I1001 10:47:29.516964 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-grcxg_508d3925-1e4f-4775-b115-1e4aeccbf260/cert-manager-controller/0.log" Oct 01 10:47:29 crc kubenswrapper[4837]: I1001 10:47:29.721564 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-xxpxb_262862ea-9ee9-4013-85ca-f8d8a196cd00/cert-manager-cainjector/0.log" Oct 01 10:47:29 crc kubenswrapper[4837]: I1001 10:47:29.875836 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-ktdtr_8b345b2c-2564-4049-94b8-0ddc18b8bd74/cert-manager-webhook/0.log" Oct 01 10:47:41 crc kubenswrapper[4837]: I1001 10:47:41.816047 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:47:41 crc kubenswrapper[4837]: E1001 10:47:41.816846 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.274389 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-lbjd2_f00c5dd2-4cb3-4934-86d1-7fcd021441e1/nmstate-console-plugin/0.log" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.457465 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-pwg46_2e35d01a-91a7-4d3b-b9de-c688b3a7f10e/nmstate-handler/0.log" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.494371 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-47mbv_b62aa771-d06e-4b59-ab29-ad40a92198ee/kube-rbac-proxy/0.log" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.576664 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-47mbv_b62aa771-d06e-4b59-ab29-ad40a92198ee/nmstate-metrics/0.log" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.790105 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-bcn95_71c13aee-6f73-4172-9a40-805d4b0c45bc/nmstate-webhook/0.log" Oct 01 10:47:43 crc kubenswrapper[4837]: I1001 10:47:43.830933 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-v5nzr_8898e3aa-b275-46de-92f9-1e1dd904d9bc/nmstate-operator/0.log" Oct 01 10:47:55 crc kubenswrapper[4837]: I1001 10:47:55.815770 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:47:55 crc kubenswrapper[4837]: E1001 10:47:55.817558 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.103642 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-hdqz9_116f7e17-f8af-4518-af7d-f2ad1ab06acf/kube-rbac-proxy/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.386332 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-frr-files/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.665882 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-hdqz9_116f7e17-f8af-4518-af7d-f2ad1ab06acf/controller/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.682368 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-reloader/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.751621 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-frr-files/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.785706 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-metrics/0.log" Oct 01 10:47:59 crc kubenswrapper[4837]: I1001 10:47:59.873558 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-reloader/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.058725 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-frr-files/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.061019 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-metrics/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.132870 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-reloader/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.146558 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-metrics/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.304022 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-reloader/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.353606 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-frr-files/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.369295 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/cp-metrics/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.423976 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/controller/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.612255 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/kube-rbac-proxy/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.614723 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/frr-metrics/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.843776 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/kube-rbac-proxy-frr/0.log" Oct 01 10:48:00 crc kubenswrapper[4837]: I1001 10:48:00.886573 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/reloader/0.log" Oct 01 10:48:01 crc kubenswrapper[4837]: I1001 10:48:01.259883 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-s9dpr_78113597-c462-4e42-8439-6d8b74465dbf/frr-k8s-webhook-server/0.log" Oct 01 10:48:01 crc kubenswrapper[4837]: I1001 10:48:01.339093 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75b5d968bd-ggvqv_b06a7555-980f-454d-8459-7385bcdd4471/manager/0.log" Oct 01 10:48:01 crc kubenswrapper[4837]: I1001 10:48:01.578523 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-876c594b8-snjgm_a2da66dd-d078-41cb-ae82-0aa30cdc04cd/webhook-server/0.log" Oct 01 10:48:01 crc kubenswrapper[4837]: I1001 10:48:01.900090 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-c4jff_96e8ea69-4a48-4923-94c7-cd3fb01b6f69/kube-rbac-proxy/0.log" Oct 01 10:48:02 crc kubenswrapper[4837]: I1001 10:48:02.797172 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-c4jff_96e8ea69-4a48-4923-94c7-cd3fb01b6f69/speaker/0.log" Oct 01 10:48:04 crc kubenswrapper[4837]: I1001 10:48:04.024209 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qr2xb_25ceb3dd-39ef-4039-b923-da36fb01ff05/frr/0.log" Oct 01 10:48:08 crc kubenswrapper[4837]: I1001 10:48:08.816394 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:48:08 crc kubenswrapper[4837]: E1001 10:48:08.818435 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:48:18 crc kubenswrapper[4837]: I1001 10:48:18.469410 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/util/0.log" Oct 01 10:48:18 crc kubenswrapper[4837]: I1001 10:48:18.646212 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/util/0.log" Oct 01 10:48:18 crc kubenswrapper[4837]: I1001 10:48:18.739750 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/pull/0.log" Oct 01 10:48:18 crc kubenswrapper[4837]: I1001 10:48:18.744748 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/pull/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.014568 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/pull/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.019870 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/util/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.050937 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb696tnr4_f8c13b9c-27b5-44bb-b820-f1e4dca3e302/extract/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.175438 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/util/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.404559 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/pull/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.472966 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/util/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.599580 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/pull/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.658726 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/pull/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.666760 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/util/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.740023 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bchdnqf_bae36cbc-90d3-46bc-a06b-4dc3a17721c9/extract/0.log" Oct 01 10:48:19 crc kubenswrapper[4837]: I1001 10:48:19.941168 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/util/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.078367 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/util/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.091479 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/pull/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.114353 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/pull/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.320964 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/util/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.325350 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/pull/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.390098 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dvx2nc_a9cdebd9-f254-406b-9e2a-775993bc8b11/extract/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.515829 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-utilities/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.728325 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-content/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.759773 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-content/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.770850 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-utilities/0.log" Oct 01 10:48:20 crc kubenswrapper[4837]: I1001 10:48:20.996569 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-content/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.114330 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/extract-utilities/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.310244 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-utilities/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.664854 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-utilities/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.668748 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-content/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.676433 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-content/0.log" Oct 01 10:48:21 crc kubenswrapper[4837]: I1001 10:48:21.817048 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:48:21 crc kubenswrapper[4837]: E1001 10:48:21.817264 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.110898 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-utilities/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.122459 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/extract-content/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.491722 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/util/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.606633 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gdh98_22813b4c-26e4-41f6-ae6a-61a735db8ec5/registry-server/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.699630 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/util/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.856757 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/pull/0.log" Oct 01 10:48:22 crc kubenswrapper[4837]: I1001 10:48:22.908894 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/pull/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.203921 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/util/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.217299 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/pull/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.252892 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d964gbll_c2c30e8f-8087-40d0-845a-4ea48e0547c6/extract/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.500758 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-t7lps_03ef9a21-4a00-49ac-9891-ad16e54b9d84/marketplace-operator/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.657944 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-utilities/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.799721 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-content/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.811843 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-utilities/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.908035 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-content/0.log" Oct 01 10:48:23 crc kubenswrapper[4837]: I1001 10:48:23.935338 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ln2gs_82ec76e8-dee7-47ad-be82-6eca642819b3/registry-server/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.109778 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-content/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.198911 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/extract-utilities/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.234593 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-utilities/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.368755 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-utilities/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.482612 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-content/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.528754 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-content/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.592784 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-bh2pb_619e4df5-ec87-4ba4-8d5a-1f2c33f028fc/registry-server/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.707136 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-content/0.log" Oct 01 10:48:24 crc kubenswrapper[4837]: I1001 10:48:24.766249 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/extract-utilities/0.log" Oct 01 10:48:26 crc kubenswrapper[4837]: I1001 10:48:26.109045 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lljk9_09d2ae22-a924-47b6-8658-eb6dbabdfa46/registry-server/0.log" Oct 01 10:48:36 crc kubenswrapper[4837]: I1001 10:48:36.815540 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:48:36 crc kubenswrapper[4837]: E1001 10:48:36.816357 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:48:38 crc kubenswrapper[4837]: I1001 10:48:38.405953 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-vh2dz_90905d8d-f989-49a4-9f32-2648922f050b/prometheus-operator/0.log" Oct 01 10:48:38 crc kubenswrapper[4837]: I1001 10:48:38.596397 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-75cf88f45d-mzbrz_a2fc8670-131c-486d-800e-57b543426d8c/prometheus-operator-admission-webhook/0.log" Oct 01 10:48:38 crc kubenswrapper[4837]: I1001 10:48:38.744879 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-75cf88f45d-s2kgk_10948123-507c-4b8b-9333-0a29d502537e/prometheus-operator-admission-webhook/0.log" Oct 01 10:48:38 crc kubenswrapper[4837]: I1001 10:48:38.821076 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-q7w4h_2881b32b-58f6-4790-9011-82e66406b813/operator/0.log" Oct 01 10:48:39 crc kubenswrapper[4837]: I1001 10:48:39.112616 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-jmfpx_ff0addb2-dc6a-4838-ac75-e5cbcd7d33eb/perses-operator/0.log" Oct 01 10:48:51 crc kubenswrapper[4837]: I1001 10:48:51.816229 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:48:51 crc kubenswrapper[4837]: E1001 10:48:51.817088 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:49:04 crc kubenswrapper[4837]: I1001 10:49:04.817146 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:49:04 crc kubenswrapper[4837]: E1001 10:49:04.817817 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:49:17 crc kubenswrapper[4837]: I1001 10:49:17.816097 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:49:17 crc kubenswrapper[4837]: E1001 10:49:17.817045 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:49:30 crc kubenswrapper[4837]: I1001 10:49:30.821265 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:49:30 crc kubenswrapper[4837]: E1001 10:49:30.823661 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:49:41 crc kubenswrapper[4837]: I1001 10:49:41.816236 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:49:41 crc kubenswrapper[4837]: E1001 10:49:41.816887 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:49:54 crc kubenswrapper[4837]: I1001 10:49:54.816192 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:49:54 crc kubenswrapper[4837]: E1001 10:49:54.818995 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:50:06 crc kubenswrapper[4837]: I1001 10:50:06.815831 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:50:06 crc kubenswrapper[4837]: E1001 10:50:06.816476 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:50:20 crc kubenswrapper[4837]: I1001 10:50:20.816814 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:50:20 crc kubenswrapper[4837]: E1001 10:50:20.818726 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:50:33 crc kubenswrapper[4837]: I1001 10:50:33.824977 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:50:33 crc kubenswrapper[4837]: E1001 10:50:33.825932 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:50:47 crc kubenswrapper[4837]: I1001 10:50:47.818258 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:50:47 crc kubenswrapper[4837]: E1001 10:50:47.818996 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:50:59 crc kubenswrapper[4837]: I1001 10:50:59.816830 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:50:59 crc kubenswrapper[4837]: E1001 10:50:59.819024 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:51:10 crc kubenswrapper[4837]: I1001 10:51:10.817247 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:51:10 crc kubenswrapper[4837]: E1001 10:51:10.818306 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:51:22 crc kubenswrapper[4837]: I1001 10:51:22.815765 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:51:22 crc kubenswrapper[4837]: E1001 10:51:22.816515 4837 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lpk6v_openshift-machine-config-operator(6e12bb16-7807-437f-927c-71dc8d0a27ec)\"" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" Oct 01 10:51:37 crc kubenswrapper[4837]: I1001 10:51:37.818863 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:51:39 crc kubenswrapper[4837]: I1001 10:51:39.187916 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"c1c876fa5d5ef31bee164c0c846608b0c766d06bb6e880816cc5d3c79517a04b"} Oct 01 10:52:27 crc kubenswrapper[4837]: I1001 10:52:27.110281 4837 scope.go:117] "RemoveContainer" containerID="99289b4707c62fc27651295fbf0ad8dad630b570b04a6a3baf6fc36503a93c5a" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.516600 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:52:54 crc kubenswrapper[4837]: E1001 10:52:54.517740 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="extract-content" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.517761 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="extract-content" Oct 01 10:52:54 crc kubenswrapper[4837]: E1001 10:52:54.517780 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="registry-server" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.517788 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="registry-server" Oct 01 10:52:54 crc kubenswrapper[4837]: E1001 10:52:54.517832 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="extract-utilities" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.517842 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="extract-utilities" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.518091 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="9156916c-49ea-47af-88e4-aa7941667362" containerName="registry-server" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.520227 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.531107 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.554570 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.554640 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.554768 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xbgl\" (UniqueName: \"kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.657464 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xbgl\" (UniqueName: \"kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.657659 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.657754 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.658409 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.658436 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.680992 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xbgl\" (UniqueName: \"kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl\") pod \"certified-operators-s9wjv\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:54 crc kubenswrapper[4837]: I1001 10:52:54.868634 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:52:55 crc kubenswrapper[4837]: I1001 10:52:55.641982 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:52:56 crc kubenswrapper[4837]: I1001 10:52:56.013383 4837 generic.go:334] "Generic (PLEG): container finished" podID="7126f90a-7849-4ae7-a97e-76077f496be2" containerID="693189d63276edc821c3e49dad65a2a994ba6de03cf57daf7186b46029d8f2b0" exitCode=0 Oct 01 10:52:56 crc kubenswrapper[4837]: I1001 10:52:56.013531 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerDied","Data":"693189d63276edc821c3e49dad65a2a994ba6de03cf57daf7186b46029d8f2b0"} Oct 01 10:52:56 crc kubenswrapper[4837]: I1001 10:52:56.013739 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerStarted","Data":"3bc79b3641952c79929020ed86063880f607481a91508e326bea8df9ce404035"} Oct 01 10:52:56 crc kubenswrapper[4837]: I1001 10:52:56.015807 4837 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:52:57 crc kubenswrapper[4837]: I1001 10:52:57.029859 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerStarted","Data":"b88bac9e281e97b0c7a8036981cf78e70d038b0c15d0d8375676484366ee33b8"} Oct 01 10:52:59 crc kubenswrapper[4837]: I1001 10:52:59.054870 4837 generic.go:334] "Generic (PLEG): container finished" podID="7126f90a-7849-4ae7-a97e-76077f496be2" containerID="b88bac9e281e97b0c7a8036981cf78e70d038b0c15d0d8375676484366ee33b8" exitCode=0 Oct 01 10:52:59 crc kubenswrapper[4837]: I1001 10:52:59.054910 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerDied","Data":"b88bac9e281e97b0c7a8036981cf78e70d038b0c15d0d8375676484366ee33b8"} Oct 01 10:53:00 crc kubenswrapper[4837]: I1001 10:53:00.069914 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerStarted","Data":"d9c6d0609943941e38100ac63626cdb8951cd70dba543cde00111258ce648b1d"} Oct 01 10:53:00 crc kubenswrapper[4837]: I1001 10:53:00.111093 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s9wjv" podStartSLOduration=2.652180496 podStartE2EDuration="6.111068075s" podCreationTimestamp="2025-10-01 10:52:54 +0000 UTC" firstStartedPulling="2025-10-01 10:52:56.015406043 +0000 UTC m=+13632.857013508" lastFinishedPulling="2025-10-01 10:52:59.474293632 +0000 UTC m=+13636.315901087" observedRunningTime="2025-10-01 10:53:00.096637708 +0000 UTC m=+13636.938245193" watchObservedRunningTime="2025-10-01 10:53:00.111068075 +0000 UTC m=+13636.952675530" Oct 01 10:53:04 crc kubenswrapper[4837]: I1001 10:53:04.870043 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:04 crc kubenswrapper[4837]: I1001 10:53:04.870387 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:04 crc kubenswrapper[4837]: I1001 10:53:04.936079 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:05 crc kubenswrapper[4837]: I1001 10:53:05.194468 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:05 crc kubenswrapper[4837]: I1001 10:53:05.275146 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:53:07 crc kubenswrapper[4837]: I1001 10:53:07.147674 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s9wjv" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="registry-server" containerID="cri-o://d9c6d0609943941e38100ac63626cdb8951cd70dba543cde00111258ce648b1d" gracePeriod=2 Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.161342 4837 generic.go:334] "Generic (PLEG): container finished" podID="7126f90a-7849-4ae7-a97e-76077f496be2" containerID="d9c6d0609943941e38100ac63626cdb8951cd70dba543cde00111258ce648b1d" exitCode=0 Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.161985 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerDied","Data":"d9c6d0609943941e38100ac63626cdb8951cd70dba543cde00111258ce648b1d"} Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.834184 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.960763 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities\") pod \"7126f90a-7849-4ae7-a97e-76077f496be2\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.960850 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content\") pod \"7126f90a-7849-4ae7-a97e-76077f496be2\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.960956 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xbgl\" (UniqueName: \"kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl\") pod \"7126f90a-7849-4ae7-a97e-76077f496be2\" (UID: \"7126f90a-7849-4ae7-a97e-76077f496be2\") " Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.961876 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities" (OuterVolumeSpecName: "utilities") pod "7126f90a-7849-4ae7-a97e-76077f496be2" (UID: "7126f90a-7849-4ae7-a97e-76077f496be2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:53:08 crc kubenswrapper[4837]: I1001 10:53:08.974426 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl" (OuterVolumeSpecName: "kube-api-access-4xbgl") pod "7126f90a-7849-4ae7-a97e-76077f496be2" (UID: "7126f90a-7849-4ae7-a97e-76077f496be2"). InnerVolumeSpecName "kube-api-access-4xbgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.007104 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7126f90a-7849-4ae7-a97e-76077f496be2" (UID: "7126f90a-7849-4ae7-a97e-76077f496be2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.063598 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xbgl\" (UniqueName: \"kubernetes.io/projected/7126f90a-7849-4ae7-a97e-76077f496be2-kube-api-access-4xbgl\") on node \"crc\" DevicePath \"\"" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.063636 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.063649 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7126f90a-7849-4ae7-a97e-76077f496be2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.175043 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9wjv" event={"ID":"7126f90a-7849-4ae7-a97e-76077f496be2","Type":"ContainerDied","Data":"3bc79b3641952c79929020ed86063880f607481a91508e326bea8df9ce404035"} Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.175098 4837 scope.go:117] "RemoveContainer" containerID="d9c6d0609943941e38100ac63626cdb8951cd70dba543cde00111258ce648b1d" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.175434 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9wjv" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.218043 4837 scope.go:117] "RemoveContainer" containerID="b88bac9e281e97b0c7a8036981cf78e70d038b0c15d0d8375676484366ee33b8" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.219823 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.236793 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s9wjv"] Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.240881 4837 scope.go:117] "RemoveContainer" containerID="693189d63276edc821c3e49dad65a2a994ba6de03cf57daf7186b46029d8f2b0" Oct 01 10:53:09 crc kubenswrapper[4837]: I1001 10:53:09.829315 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" path="/var/lib/kubelet/pods/7126f90a-7849-4ae7-a97e-76077f496be2/volumes" Oct 01 10:53:39 crc kubenswrapper[4837]: I1001 10:53:39.480506 4837 generic.go:334] "Generic (PLEG): container finished" podID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerID="ef8f14431fb346eb25686538db0ee7e102dd8a3b9f31148c0a0857c1012bf857" exitCode=0 Oct 01 10:53:39 crc kubenswrapper[4837]: I1001 10:53:39.480588 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" event={"ID":"832b5f04-3558-43f1-84b3-abb7a96ba682","Type":"ContainerDied","Data":"ef8f14431fb346eb25686538db0ee7e102dd8a3b9f31148c0a0857c1012bf857"} Oct 01 10:53:39 crc kubenswrapper[4837]: I1001 10:53:39.481664 4837 scope.go:117] "RemoveContainer" containerID="ef8f14431fb346eb25686538db0ee7e102dd8a3b9f31148c0a0857c1012bf857" Oct 01 10:53:40 crc kubenswrapper[4837]: I1001 10:53:40.177162 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ntfc2_must-gather-6rdm2_832b5f04-3558-43f1-84b3-abb7a96ba682/gather/0.log" Oct 01 10:53:53 crc kubenswrapper[4837]: I1001 10:53:53.079645 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:53:53 crc kubenswrapper[4837]: I1001 10:53:53.080103 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:53:58 crc kubenswrapper[4837]: I1001 10:53:58.633153 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ntfc2/must-gather-6rdm2"] Oct 01 10:53:58 crc kubenswrapper[4837]: I1001 10:53:58.634208 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="copy" containerID="cri-o://ac02b73ec4edc289f196cc7ffd679bed047ad706f196286c6dc1072fff34e61a" gracePeriod=2 Oct 01 10:53:58 crc kubenswrapper[4837]: I1001 10:53:58.645746 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ntfc2/must-gather-6rdm2"] Oct 01 10:53:59 crc kubenswrapper[4837]: I1001 10:53:59.694107 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ntfc2_must-gather-6rdm2_832b5f04-3558-43f1-84b3-abb7a96ba682/copy/0.log" Oct 01 10:53:59 crc kubenswrapper[4837]: I1001 10:53:59.694903 4837 generic.go:334] "Generic (PLEG): container finished" podID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerID="ac02b73ec4edc289f196cc7ffd679bed047ad706f196286c6dc1072fff34e61a" exitCode=143 Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.106579 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ntfc2_must-gather-6rdm2_832b5f04-3558-43f1-84b3-abb7a96ba682/copy/0.log" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.107332 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.214385 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output\") pod \"832b5f04-3558-43f1-84b3-abb7a96ba682\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.214436 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccbm5\" (UniqueName: \"kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5\") pod \"832b5f04-3558-43f1-84b3-abb7a96ba682\" (UID: \"832b5f04-3558-43f1-84b3-abb7a96ba682\") " Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.220138 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5" (OuterVolumeSpecName: "kube-api-access-ccbm5") pod "832b5f04-3558-43f1-84b3-abb7a96ba682" (UID: "832b5f04-3558-43f1-84b3-abb7a96ba682"). InnerVolumeSpecName "kube-api-access-ccbm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.316872 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccbm5\" (UniqueName: \"kubernetes.io/projected/832b5f04-3558-43f1-84b3-abb7a96ba682-kube-api-access-ccbm5\") on node \"crc\" DevicePath \"\"" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.525237 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "832b5f04-3558-43f1-84b3-abb7a96ba682" (UID: "832b5f04-3558-43f1-84b3-abb7a96ba682"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.622844 4837 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/832b5f04-3558-43f1-84b3-abb7a96ba682-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.708899 4837 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ntfc2_must-gather-6rdm2_832b5f04-3558-43f1-84b3-abb7a96ba682/copy/0.log" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.709419 4837 scope.go:117] "RemoveContainer" containerID="ac02b73ec4edc289f196cc7ffd679bed047ad706f196286c6dc1072fff34e61a" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.709475 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ntfc2/must-gather-6rdm2" Oct 01 10:54:00 crc kubenswrapper[4837]: I1001 10:54:00.740774 4837 scope.go:117] "RemoveContainer" containerID="ef8f14431fb346eb25686538db0ee7e102dd8a3b9f31148c0a0857c1012bf857" Oct 01 10:54:01 crc kubenswrapper[4837]: I1001 10:54:01.846501 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" path="/var/lib/kubelet/pods/832b5f04-3558-43f1-84b3-abb7a96ba682/volumes" Oct 01 10:54:23 crc kubenswrapper[4837]: I1001 10:54:23.079015 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:54:23 crc kubenswrapper[4837]: I1001 10:54:23.080463 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.703583 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:54:52 crc kubenswrapper[4837]: E1001 10:54:52.704941 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="extract-content" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.704961 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="extract-content" Oct 01 10:54:52 crc kubenswrapper[4837]: E1001 10:54:52.704981 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="extract-utilities" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.704989 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="extract-utilities" Oct 01 10:54:52 crc kubenswrapper[4837]: E1001 10:54:52.705037 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="registry-server" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705045 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="registry-server" Oct 01 10:54:52 crc kubenswrapper[4837]: E1001 10:54:52.705055 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="gather" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705063 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="gather" Oct 01 10:54:52 crc kubenswrapper[4837]: E1001 10:54:52.705078 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="copy" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705085 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="copy" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705326 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="copy" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705354 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="832b5f04-3558-43f1-84b3-abb7a96ba682" containerName="gather" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.705377 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="7126f90a-7849-4ae7-a97e-76077f496be2" containerName="registry-server" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.707316 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.714350 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.857523 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.857687 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.857753 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7zx\" (UniqueName: \"kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.959754 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.959863 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7zx\" (UniqueName: \"kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.959933 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.960263 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.960386 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:52 crc kubenswrapper[4837]: I1001 10:54:52.977465 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7zx\" (UniqueName: \"kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx\") pod \"community-operators-lc8g7\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.041591 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.079742 4837 patch_prober.go:28] interesting pod/machine-config-daemon-lpk6v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.080037 4837 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.080082 4837 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.080861 4837 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1c876fa5d5ef31bee164c0c846608b0c766d06bb6e880816cc5d3c79517a04b"} pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.080916 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" podUID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerName="machine-config-daemon" containerID="cri-o://c1c876fa5d5ef31bee164c0c846608b0c766d06bb6e880816cc5d3c79517a04b" gracePeriod=600 Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.264388 4837 generic.go:334] "Generic (PLEG): container finished" podID="6e12bb16-7807-437f-927c-71dc8d0a27ec" containerID="c1c876fa5d5ef31bee164c0c846608b0c766d06bb6e880816cc5d3c79517a04b" exitCode=0 Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.264434 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerDied","Data":"c1c876fa5d5ef31bee164c0c846608b0c766d06bb6e880816cc5d3c79517a04b"} Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.264502 4837 scope.go:117] "RemoveContainer" containerID="fc4159c5b281a20ed7ca828b5e4baa2c887e90610797cbcbffc2eb2195ebf6aa" Oct 01 10:54:53 crc kubenswrapper[4837]: I1001 10:54:53.919599 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:54:53 crc kubenswrapper[4837]: W1001 10:54:53.930332 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aafb726_d999_4f0d_ae8d_530477d93267.slice/crio-47b62fac9e577e41bb476bceafa71f38d398a0bfd2cdf1b8cc0bee11980db8fd WatchSource:0}: Error finding container 47b62fac9e577e41bb476bceafa71f38d398a0bfd2cdf1b8cc0bee11980db8fd: Status 404 returned error can't find the container with id 47b62fac9e577e41bb476bceafa71f38d398a0bfd2cdf1b8cc0bee11980db8fd Oct 01 10:54:54 crc kubenswrapper[4837]: I1001 10:54:54.276494 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lpk6v" event={"ID":"6e12bb16-7807-437f-927c-71dc8d0a27ec","Type":"ContainerStarted","Data":"f0a8eba56fdcaf867b944a048a6c752369ea2a66c760d8e901dca432b29dffa8"} Oct 01 10:54:54 crc kubenswrapper[4837]: I1001 10:54:54.278551 4837 generic.go:334] "Generic (PLEG): container finished" podID="0aafb726-d999-4f0d-ae8d-530477d93267" containerID="6ad761b4a89666f2025c5ad5f1d3353ecaaf504362313e6073e6ee048c902a1a" exitCode=0 Oct 01 10:54:54 crc kubenswrapper[4837]: I1001 10:54:54.278613 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerDied","Data":"6ad761b4a89666f2025c5ad5f1d3353ecaaf504362313e6073e6ee048c902a1a"} Oct 01 10:54:54 crc kubenswrapper[4837]: I1001 10:54:54.278654 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerStarted","Data":"47b62fac9e577e41bb476bceafa71f38d398a0bfd2cdf1b8cc0bee11980db8fd"} Oct 01 10:54:55 crc kubenswrapper[4837]: I1001 10:54:55.293831 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerStarted","Data":"061151053f2f6a55b4c962aea762aeed567c1befffb76a3e0f19398c43539140"} Oct 01 10:54:58 crc kubenswrapper[4837]: I1001 10:54:58.335621 4837 generic.go:334] "Generic (PLEG): container finished" podID="0aafb726-d999-4f0d-ae8d-530477d93267" containerID="061151053f2f6a55b4c962aea762aeed567c1befffb76a3e0f19398c43539140" exitCode=0 Oct 01 10:54:58 crc kubenswrapper[4837]: I1001 10:54:58.335714 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerDied","Data":"061151053f2f6a55b4c962aea762aeed567c1befffb76a3e0f19398c43539140"} Oct 01 10:54:59 crc kubenswrapper[4837]: I1001 10:54:59.365424 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerStarted","Data":"d9d9ba23ae122ae6dc91930a5a016bcc4b3a8c03b8e7038fb348cd0565b2f6ff"} Oct 01 10:54:59 crc kubenswrapper[4837]: I1001 10:54:59.398654 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lc8g7" podStartSLOduration=2.912116182 podStartE2EDuration="7.398633576s" podCreationTimestamp="2025-10-01 10:54:52 +0000 UTC" firstStartedPulling="2025-10-01 10:54:54.280554507 +0000 UTC m=+13751.122161962" lastFinishedPulling="2025-10-01 10:54:58.767071901 +0000 UTC m=+13755.608679356" observedRunningTime="2025-10-01 10:54:59.390074574 +0000 UTC m=+13756.231682049" watchObservedRunningTime="2025-10-01 10:54:59.398633576 +0000 UTC m=+13756.240241041" Oct 01 10:55:03 crc kubenswrapper[4837]: I1001 10:55:03.041800 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:03 crc kubenswrapper[4837]: I1001 10:55:03.042441 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:03 crc kubenswrapper[4837]: I1001 10:55:03.104314 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:13 crc kubenswrapper[4837]: I1001 10:55:13.161114 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:13 crc kubenswrapper[4837]: I1001 10:55:13.233424 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:55:13 crc kubenswrapper[4837]: I1001 10:55:13.491079 4837 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lc8g7" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="registry-server" containerID="cri-o://d9d9ba23ae122ae6dc91930a5a016bcc4b3a8c03b8e7038fb348cd0565b2f6ff" gracePeriod=2 Oct 01 10:55:14 crc kubenswrapper[4837]: I1001 10:55:14.518579 4837 generic.go:334] "Generic (PLEG): container finished" podID="0aafb726-d999-4f0d-ae8d-530477d93267" containerID="d9d9ba23ae122ae6dc91930a5a016bcc4b3a8c03b8e7038fb348cd0565b2f6ff" exitCode=0 Oct 01 10:55:14 crc kubenswrapper[4837]: I1001 10:55:14.518627 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerDied","Data":"d9d9ba23ae122ae6dc91930a5a016bcc4b3a8c03b8e7038fb348cd0565b2f6ff"} Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.252558 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.314361 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h7zx\" (UniqueName: \"kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx\") pod \"0aafb726-d999-4f0d-ae8d-530477d93267\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.314416 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities\") pod \"0aafb726-d999-4f0d-ae8d-530477d93267\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.314444 4837 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content\") pod \"0aafb726-d999-4f0d-ae8d-530477d93267\" (UID: \"0aafb726-d999-4f0d-ae8d-530477d93267\") " Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.316740 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities" (OuterVolumeSpecName: "utilities") pod "0aafb726-d999-4f0d-ae8d-530477d93267" (UID: "0aafb726-d999-4f0d-ae8d-530477d93267"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.326778 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx" (OuterVolumeSpecName: "kube-api-access-5h7zx") pod "0aafb726-d999-4f0d-ae8d-530477d93267" (UID: "0aafb726-d999-4f0d-ae8d-530477d93267"). InnerVolumeSpecName "kube-api-access-5h7zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.378423 4837 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0aafb726-d999-4f0d-ae8d-530477d93267" (UID: "0aafb726-d999-4f0d-ae8d-530477d93267"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.417010 4837 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h7zx\" (UniqueName: \"kubernetes.io/projected/0aafb726-d999-4f0d-ae8d-530477d93267-kube-api-access-5h7zx\") on node \"crc\" DevicePath \"\"" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.417045 4837 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.417057 4837 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aafb726-d999-4f0d-ae8d-530477d93267-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.530481 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lc8g7" event={"ID":"0aafb726-d999-4f0d-ae8d-530477d93267","Type":"ContainerDied","Data":"47b62fac9e577e41bb476bceafa71f38d398a0bfd2cdf1b8cc0bee11980db8fd"} Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.530535 4837 scope.go:117] "RemoveContainer" containerID="d9d9ba23ae122ae6dc91930a5a016bcc4b3a8c03b8e7038fb348cd0565b2f6ff" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.530542 4837 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lc8g7" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.553595 4837 scope.go:117] "RemoveContainer" containerID="061151053f2f6a55b4c962aea762aeed567c1befffb76a3e0f19398c43539140" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.572233 4837 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.599446 4837 scope.go:117] "RemoveContainer" containerID="6ad761b4a89666f2025c5ad5f1d3353ecaaf504362313e6073e6ee048c902a1a" Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.603294 4837 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lc8g7"] Oct 01 10:55:15 crc kubenswrapper[4837]: I1001 10:55:15.829943 4837 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" path="/var/lib/kubelet/pods/0aafb726-d999-4f0d-ae8d-530477d93267/volumes" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.938687 4837 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5jcq"] Oct 01 10:56:02 crc kubenswrapper[4837]: E1001 10:56:02.939652 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="extract-content" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.939664 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="extract-content" Oct 01 10:56:02 crc kubenswrapper[4837]: E1001 10:56:02.939676 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="registry-server" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.939682 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="registry-server" Oct 01 10:56:02 crc kubenswrapper[4837]: E1001 10:56:02.939736 4837 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="extract-utilities" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.939743 4837 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="extract-utilities" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.939957 4837 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aafb726-d999-4f0d-ae8d-530477d93267" containerName="registry-server" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.941444 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:02 crc kubenswrapper[4837]: I1001 10:56:02.960482 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5jcq"] Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.092573 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-utilities\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.092636 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfcbb\" (UniqueName: \"kubernetes.io/projected/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-kube-api-access-hfcbb\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.092947 4837 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-catalog-content\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.195424 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-utilities\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.195548 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfcbb\" (UniqueName: \"kubernetes.io/projected/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-kube-api-access-hfcbb\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.195683 4837 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-catalog-content\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.196293 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-utilities\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.196515 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-catalog-content\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.219119 4837 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfcbb\" (UniqueName: \"kubernetes.io/projected/25168520-2bc7-4b4f-9ad2-1587d93bb7a4-kube-api-access-hfcbb\") pod \"redhat-operators-g5jcq\" (UID: \"25168520-2bc7-4b4f-9ad2-1587d93bb7a4\") " pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:03 crc kubenswrapper[4837]: I1001 10:56:03.266360 4837 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:04 crc kubenswrapper[4837]: I1001 10:56:04.039142 4837 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5jcq"] Oct 01 10:56:04 crc kubenswrapper[4837]: W1001 10:56:04.043860 4837 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25168520_2bc7_4b4f_9ad2_1587d93bb7a4.slice/crio-f904b01af79991b05d9901556c8855d3e5dc961aefcc063ceea389fa89d467ab WatchSource:0}: Error finding container f904b01af79991b05d9901556c8855d3e5dc961aefcc063ceea389fa89d467ab: Status 404 returned error can't find the container with id f904b01af79991b05d9901556c8855d3e5dc961aefcc063ceea389fa89d467ab Oct 01 10:56:04 crc kubenswrapper[4837]: I1001 10:56:04.161688 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5jcq" event={"ID":"25168520-2bc7-4b4f-9ad2-1587d93bb7a4","Type":"ContainerStarted","Data":"f904b01af79991b05d9901556c8855d3e5dc961aefcc063ceea389fa89d467ab"} Oct 01 10:56:05 crc kubenswrapper[4837]: I1001 10:56:05.178188 4837 generic.go:334] "Generic (PLEG): container finished" podID="25168520-2bc7-4b4f-9ad2-1587d93bb7a4" containerID="823aa1ed011b83111bde6706cdce3f076d6e2d77d93ceea11345a50d21de4bb9" exitCode=0 Oct 01 10:56:05 crc kubenswrapper[4837]: I1001 10:56:05.178272 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5jcq" event={"ID":"25168520-2bc7-4b4f-9ad2-1587d93bb7a4","Type":"ContainerDied","Data":"823aa1ed011b83111bde6706cdce3f076d6e2d77d93ceea11345a50d21de4bb9"} Oct 01 10:56:07 crc kubenswrapper[4837]: I1001 10:56:07.202518 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5jcq" event={"ID":"25168520-2bc7-4b4f-9ad2-1587d93bb7a4","Type":"ContainerStarted","Data":"ffc4db25c6f67db3696a230fea7ff54d517ef8194373c9f82effdeae9bbbf3aa"} Oct 01 10:56:10 crc kubenswrapper[4837]: I1001 10:56:10.244319 4837 generic.go:334] "Generic (PLEG): container finished" podID="25168520-2bc7-4b4f-9ad2-1587d93bb7a4" containerID="ffc4db25c6f67db3696a230fea7ff54d517ef8194373c9f82effdeae9bbbf3aa" exitCode=0 Oct 01 10:56:10 crc kubenswrapper[4837]: I1001 10:56:10.244442 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5jcq" event={"ID":"25168520-2bc7-4b4f-9ad2-1587d93bb7a4","Type":"ContainerDied","Data":"ffc4db25c6f67db3696a230fea7ff54d517ef8194373c9f82effdeae9bbbf3aa"} Oct 01 10:56:11 crc kubenswrapper[4837]: I1001 10:56:11.259340 4837 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5jcq" event={"ID":"25168520-2bc7-4b4f-9ad2-1587d93bb7a4","Type":"ContainerStarted","Data":"56183a3c637f2ac14c2364da260a2ecb3aadd681f1d538612f4f76f666bb5663"} Oct 01 10:56:11 crc kubenswrapper[4837]: I1001 10:56:11.287652 4837 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5jcq" podStartSLOduration=3.727223221 podStartE2EDuration="9.28762882s" podCreationTimestamp="2025-10-01 10:56:02 +0000 UTC" firstStartedPulling="2025-10-01 10:56:05.181554004 +0000 UTC m=+13822.023161489" lastFinishedPulling="2025-10-01 10:56:10.741959623 +0000 UTC m=+13827.583567088" observedRunningTime="2025-10-01 10:56:11.283056627 +0000 UTC m=+13828.124664092" watchObservedRunningTime="2025-10-01 10:56:11.28762882 +0000 UTC m=+13828.129236285" Oct 01 10:56:13 crc kubenswrapper[4837]: I1001 10:56:13.266798 4837 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5jcq" Oct 01 10:56:13 crc kubenswrapper[4837]: I1001 10:56:13.267343 4837 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5jcq" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067204325024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067204325017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067150363016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067150363015463 5ustar corecore